Example

Diagrammatic Example of an Encoder-Decoder Model Trained with a Denoising Autoencoding Objective

This diagram provides an example of training an encoder-decoder model using a denoising autoencoding objective. The process involves several key steps: 1. Corrupted Input to Encoder: The encoder receives a corrupted version of a sentence where some tokens are masked, for instance, [CLS] The puppies are [MASK] outside [MASK] house .. 2. Sequence Reconstruction by Decoder: The encoder generates a hidden state representation of the input, which is then passed to the decoder. The decoder's task is to reconstruct the original, uncorrupted sentence, ⟨s⟩ The puppies are frolicking outside the house ., in an autoregressive manner. 3. Sequence-Level Loss Calculation: To train the model, a loss is calculated over the entire output sequence by accumulating the losses of all tokens, as in standard language modeling. This involves comparing the decoder's generated output with the ground-truth sequence, and the resulting error signal is used to update the model's parameters.

Image 0

0

1

Updated 2026-04-16

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences