Learn Before
Example

Example of Simulating Standard Language Modeling via Masking

To illustrate how masking can simulate standard language modeling, consider an input sequence where every token is replaced by a mask: [CLS] [MASK] [MASK] [MASK] [MASK] [MASK] [MASK] [MASK] [MASK]. The model is then trained to predict and generate the entire corresponding sequence, such as <s> The puppies are frolicking outside the house ., effectively performing autoregressive text generation from an entirely masked context.

0

1

Updated 2026-04-16

Contributors are:

Who are from:

Tags

Foundations of Large Language Models

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences