Improving a Flawed Evaluation Prompt
A developer is using the following prompt to have a language model check a student's math homework:
Problem: What is 5 * (3 + 2)? Student's Answer: 25. Is the student's answer correct? If not, what is the right answer?
Analyze this prompt. Identify two key weaknesses in its structure for reliably guiding the model to both evaluate the student's work and provide a high-quality correction. Then, rewrite the prompt to be more effective.
0
1
Tags
Ch.3 Prompting - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
A developer is designing a two-step process for a language model to solve complex logic puzzles. In the first step, the model provides an initial solution. For the second step, the developer wants to prompt the model to critically review and improve its own initial answer. Which of the following prompts for the second step is best designed to encourage self-correction and produce a more accurate final solution?
Improving a Flawed Evaluation Prompt
Automating Homework Grading with an LLM