Evaluating Model Correction Strategies
A user asks a language model: 'What is the total number of days in the first six months of a non-leap year (January to June)?' The model incorrectly responds: 'There are 180 days.' The user wants to guide the model to the correct answer. Consider two possible follow-up prompts:
Approach A: "That doesn't seem right. Please try calculating the total number of days in the first six months of a non-leap year again."
Approach B: "Please review your previous answer. First, list the number of days in each of the first six months. Then, sum those numbers to provide a final, corrected total."
Evaluate these two approaches. Which one is more likely to lead to a correct and reliable answer from the model? Justify your choice by explaining the difference in the reasoning process each prompt encourages.
0
1
Tags
Ch.3 Prompting - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Evaluation in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Prompt Template for Answer Evaluation and Refinement
Evaluating Model Correction Strategies
Guiding a Model to Self-Correct
You are designing a conversational interaction to help a language model solve a complex problem more reliably. Arrange the following steps into the correct logical sequence for a two-round interaction focused on self-correction.