Necessity of Post-Pre-training Alignment
The practical impossibility of collecting a pre-training dataset that is both comprehensive enough to cover all tasks and perfectly representative of human preferences means that alignment cannot be achieved through pre-training alone. Consequently, a distinct alignment stage remains an essential and critical step in the development of LLMs.
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
Necessity of Post-Pre-training Alignment
Evaluating a Pre-training-Only Strategy
A research lab proposes a new strategy to create a perfectly helpful and harmless language model. Their plan is to spend five years meticulously curating a massive dataset of text and code that only contains examples of positive, safe, and beneficial interactions. They argue that by pre-training a model exclusively on this 'perfect' dataset, no further alignment steps will be necessary. Which of the following statements identifies the most critical flaw in this strategy's approach to alignment?
Critiquing the 'Perfect Dataset' Hypothesis for Alignment
Learn After
Two-Step Post-Pre-training Alignment Process
A technology company claims it can create a perfectly helpful and harmless AI assistant by simply pre-training a model on an exhaustive dataset containing all books, articles, and websites ever published. They argue that such a comprehensive dataset would make any subsequent training phase to align the model's behavior unnecessary. Which of the following statements provides the most critical evaluation of this claim's primary flaw?
Rationale for Post-Training Alignment
Critique of a Pre-training-Only Approach