Learn Before
A development team is working to improve a large language model's behavior. They create two distinct datasets:
- Dataset 1: A curated set of prompts, each paired with a single, ideal, human-written response that demonstrates how to follow the prompt's instructions correctly.
- Dataset 2: A set of prompts where, for each prompt, a human evaluator has ranked several different model-generated responses from best to worst.
Which statement best analyzes the relationship between these datasets and the two fundamental approaches to model alignment?
0
1
Tags
Ch.5 Inference - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Ch.4 Alignment - Foundations of Large Language Models
Ch.2 Generative Models - Foundations of Large Language Models
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Surrogate Objectives in AI Alignment
Combined Use of Instruction and Human Preference Alignment
Differing Motivations of Instruction and Human Preference Alignment
Instruction Alignment
Human Preference Alignment via Reward Models
A development team is working to improve a large language model's behavior. They create two distinct datasets:
- Dataset 1: A curated set of prompts, each paired with a single, ideal, human-written response that demonstrates how to follow the prompt's instructions correctly.
- Dataset 2: A set of prompts where, for each prompt, a human evaluator has ranked several different model-generated responses from best to worst.
Which statement best analyzes the relationship between these datasets and the two fundamental approaches to model alignment?
Match each fundamental model alignment approach with its primary goal and typical implementation method.
Prioritizing Chatbot Alignment Strategies