Learn Before
Concept

Training Encoder-Decoder Models with a Denoising Autoencoding Objective

The denoising autoencoding objective is utilized to train encoder-decoder models by requiring them to reconstruct an original, uncorrupted sequence from a corrupted input. Operating similarly to a denoising autoencoder, the encoder processes the corrupted input—such as a sequence with masked tokens—and transforms it into a hidden representation. The decoder then uses this hidden representation to predict the original text. By learning to map a corrupted sequence to its uncorrupted counterpart, the model concurrently develops two key skills: the encoder gains the ability to comprehend the input context, while the decoder acquires the capability to generate coherent text.

Image 0

0

1

Updated 2026-05-02

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Related