Learn Before
Activity (Process)

Training Encoder-Decoder Models with Prefix Language Modeling

Encoder-decoder models can be trained directly using the Prefix Language Modeling objective. In this process, the encoder learns to build a contextual representation from a given prefix, while the decoder is trained to generate the subsequent text based on the encoder's understanding. This method is particularly effective for large-scale pre-training because it allows for the straightforward creation of a vast number of training examples from readily available unlabeled text.

Image 0

0

1

Updated 2026-04-16

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences