Attention Models: Learning to Focus
How the Attention mechanism solved the bottleneck problem in Seq2Seq models and paved the way for Transformers.
How the Attention mechanism solved the bottleneck problem in Seq2Seq models and paved the way for Transformers.