Learn Before
Evaluating a Verifier Model Design
A team is developing a system to automatically check the factual accuracy of summaries generated by a language model. They plan to train a separate 'verifier' model on a dataset of summaries, each labeled by a human expert. The team decides the verifier will only output one of two possible labels for any given summary: 'Accurate' or 'Inaccurate'.
Critique this design choice. What is the most significant advantage and the most significant disadvantage of limiting the verifier's output to only two categories?
0
1
Tags
Ch.3 Prompting - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Evaluation in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Evaluating a Verifier Model Design
A research team is developing a system to automatically flag whether a language model's one-sentence summary of a news article is factually correct or incorrect. They have a large dataset of summaries, each labeled as either 'Correct' or 'Incorrect'. If they frame this task as a supervised learning problem, what kind of model are they most likely training, and what would its output represent?
Justifying a Binary Classifier for Verification