Gentle Introduction to Global Attention for Encoder-Decoder Recurrent Neural Networks
Last Updated on August 14, 2019 The encoder-decoder model provides a pattern for using recurrent neural networks to address challenging sequence-to-sequence prediction problems such as machine translation. Attention is an extension to the encoder-decoder model that improves the performance of the approach on longer sequences. Global attention is a simplification of attention that may be easier to implement in declarative deep learning libraries like Keras and may achieve better results than the classic attention mechanism. In this post, you will […]
Read more