Concept

The Generator in Replaced Token Detection

In the Replaced Token Detection framework, the generator is a small masked language model tasked with creating a corrupted version of the original input text. Its process involves two main steps: first, it randomly masks a subset of tokens in a sequence. Second, it is trained to predict the original tokens for these masked positions. The generator then outputs a new sequence where the masked tokens have been replaced by its predictions, which may or may not match the original tokens. This altered sequence is then passed to the discriminator.

Image 0

0

1

Updated 2026-04-16

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Related