Concept

Purpose of Unchanged Tokens in BERT's MLM Strategy

In BERT's Masked Language Modeling strategy, predicting a target token that has been intentionally left unchanged in the input sequence is a relatively simple task. The purpose of this strategy is to guide the model to utilize easier, more direct evidence for its predictions, as the original token is explicitly available in the provided context.

0

1

Updated 2026-04-17

Contributors are:

Who are from:

Tags

Foundations of Large Language Models

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences