A language model is prompted to solve the math problem 'What is 7 + 8?'. To improve reliability, the model generates five different outputs using a sampling strategy: [15, 14, 15, 15, 16]. A selection process is then used to choose the final answer by identifying the candidate that minimizes the expected disagreement with the other generated candidates. Which output will be selected?
0
1
Tags
Ch.3 Prompting - Foundations of Large Language Models
Foundations of Large Language Models
Computing Sciences
Foundations of Large Language Models Course
Application in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Risk Function in Minimum Bayes Risk Decoding
Risk of an Output in Minimum Bayes Risk Decoding
A language model is prompted to solve the math problem 'What is 7 + 8?'. To improve reliability, the model generates five different outputs using a sampling strategy: [15, 14, 15, 15, 16]. A selection process is then used to choose the final answer by identifying the candidate that minimizes the expected disagreement with the other generated candidates. Which output will be selected?
Connecting Self-Consistency to a Formal Framework
Evaluating a Text Generation Strategy