What is a ‘recurrent neural network’ (RNN) and what limitation led to its replacement by transformers?

AI Fundamentals Medium

AI Fundamentals — Medium

What is a ‘recurrent neural network’ (RNN) and what limitation led to its replacement by transformers?

Key points

  • RNN maintains hidden state across steps for sequential data
  • Vanishing gradients and lack of parallelization limit RNN performance
  • Transformers address these limitations by allowing parallel processing

Ready to go further?

Related questions