Learn Before
The Alignment Problem in LLMs
The fundamental alignment problem is that the pre-training process for Large Language Models can itself introduce a variety of issues, leading to a discrepancy between the model's output and a user's intended goals. This misalignment occurs because pre-trained models, by default, may not have learned to follow instructions or adhere to implicit human values.
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
The Alignment Problem in LLMs
Ethical Challenges in LLM Alignment
Analysis of Model Response Alignment
A user asks a Large Language Model for a 'simple, healthy recipe for a quick lunch.' The model provides a clear, step-by-step recipe for a quinoa salad, includes a note about potential allergens, and suggests common ingredient substitutions. Which of the following statements best analyzes why this response demonstrates good alignment with human expectations?
Evaluating LLM Response Alignment