Attention is one of the most important ideas in the Deep Learning community. Although this mechanism is now used in various problems such as image captions and others, it was originally designed in the context of neural machine translation using Seq2Seq models. Seq2Seq model The seq2seq model is normally composed of an encoder-decoder architecture, in ...