Learn Before
Prioritizing Chatbot Alignment Strategies
A company's new customer service chatbot, built on a powerful language model, exhibits two key issues. First, when given a direct command like 'Summarize my last three orders,' it often continues the sentence with plausible but irrelevant text instead of executing the command. Second, when a user expresses frustration, the model provides factually correct but unempathetic and unhelpful replies. The development team must decide which alignment approach to prioritize for each issue. Analyze this scenario and determine which of the two fundamental alignment approaches is best suited to address the first issue, and which is best for the second. Justify your choices.
0
1
Tags
Ch.5 Inference - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Ch.4 Alignment - Foundations of Large Language Models
Ch.2 Generative Models - Foundations of Large Language Models
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Surrogate Objectives in AI Alignment
Combined Use of Instruction and Human Preference Alignment
Differing Motivations of Instruction and Human Preference Alignment
Instruction Alignment
Human Preference Alignment via Reward Models
A development team is working to improve a large language model's behavior. They create two distinct datasets:
- Dataset 1: A curated set of prompts, each paired with a single, ideal, human-written response that demonstrates how to follow the prompt's instructions correctly.
- Dataset 2: A set of prompts where, for each prompt, a human evaluator has ranked several different model-generated responses from best to worst.
Which statement best analyzes the relationship between these datasets and the two fundamental approaches to model alignment?
Match each fundamental model alignment approach with its primary goal and typical implementation method.
Prioritizing Chatbot Alignment Strategies