Concept

Encoder-decoder networks

The abstracted architecture of the twisted autoregressive generation is referred to as the encoder-decoder architecture.

It consists of three parts:

  • encoder: accepts a sequence as its input and generates a corresponding sequence of contextualized representations (hidden states);
  • context vector: is a function of the vector of contextualized representations generated from the encoder and conveys the essence of the input to the decoder;
  • decoder: takes context vector as input and generates an arbitrary length sequence of hidden states, therefore obtains a corresponding sequence of output states.

The encoder and decoder networks are typically implemented with the same architecture, often using recurrent networks, but there are some other possibilities in each part.

Image 0

0

4

Updated 2026-04-17

Tags

Data Science

Foundations of Large Language Models Course

Computing Sciences

Learn After