Learn Before
Limitations of a Unidirectional Pre-training Objective
A language model is pre-trained using an objective where it learns to predict each word in a text based only on the words that precede it. Analyze why this specific training method would be inherently disadvantaged for tasks that require filling in a blank in the middle of a sentence (e.g., 'The student opened their book to ___ the chapter.').
0
1
Tags
Data Science
Ch.1 Pre-training - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
A language model is being trained with the objective of predicting the next item in a sequence, given all the preceding items. If this model is processing the sentence 'The cat sat on the mat.', which of the following scenarios accurately represents a single step in its training process?
Choosing a Pre-training Objective for Text Generation
Limitations of a Unidirectional Pre-training Objective