Example

Example of Denoising Task with Consecutive Token Masking

An example of a denoising autoencoder task involving consecutive token masking is training a model to reconstruct a full sentence from a corrupted input. The model receives an input where adjacent tokens have been replaced by [MASK], and its objective is to generate the original, complete sequence. For instance, the input [CLS] The puppies are [MASK] outside [MASK] [MASK] . would be used to train the model to produce the target output: ⟨s⟩ The puppies are frolicking outside the house ..

0

1

Updated 2026-04-16

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Related