Learn Before
Concept

Encoder-Decoder with Transformers

The encoder-decoder architecture can also be implemented using transformers, consisting of:

  • An encoder that takes the source language input words X=x1,...,xTX = x_1, ..., x_T and maps them to an output representation Henc=h1,...,hTH^{enc} = h_1, ..., h_T; usually via N=6N = 6 stacked encoder blocks.
  • A decoder which is just like the one within the encoder-decoder RNN. However, decoder transformer block includes an extra cross-attention layer in order to attend to the source language.
Image 0

0

0

Updated 2021-12-05

Tags

Data Science

Related
Learn After