Trade-offs in Human Feedback Collection Methods
A team is developing a system to improve a language model's conversational abilities using human feedback. They are debating between two methods for data collection:
- Method A: Annotators rate each model-generated response on a scale of 1 to 7.
- Method B: Annotators are shown two responses to the same prompt and must choose which one is better.
Analyze the primary challenge the team would face in ensuring data quality with Method A, and explain why Method B is often considered a more reliable alternative for this type of task.
0
1
Tags
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Ch.4 Alignment - Foundations of Large Language Models
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Reward Model Learning in RLHF
Pairwise Comparison for Human Feedback in RLHF
Listwise Ranking for Human Feedback in RLHF
Preference Notation in Human Feedback
Pointwise Method (Rating) for Human Feedback in RLHF
Evaluating a Human Feedback Strategy
A research team is developing a system to improve a language model using feedback from a large, diverse group of non-expert annotators. The team's primary goal is to ensure the feedback data is as consistent and reliable as possible, even with minimal training for the annotators. Which of the following feedback collection strategies would best achieve this goal, and why?
Trade-offs in Human Feedback Collection Methods