Short Answer

Adapting a Bidirectional Model for a Unidirectional Task

A researcher has a language model that was trained to use both preceding and succeeding text to fill in missing words. They now want to use this model for a text generation task, where the goal is to predict the next word in a sentence, one word at a time. Describe the specific masking strategy the researcher must apply to the input sequence at each step of the generation process to force the model to only use the context it has already generated.

0

1

Updated 2025-10-02

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Application in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science