Learn Before
Structure of a Zero-Shot CoT Prompt for an Arithmetic Task
The structure of a zero-shot Chain-of-Thought prompt for an arithmetic task involves two key components. First, the user's question is stated directly. Second, an instruction is appended to trigger the model's reasoning process. For example, the question 'Q: Please calculate the average of the numbers 2, 4, and 9' would be followed by the instruction 'A: Let’s think step-by-step' to guide the model.
0
1
Tags
Ch.3 Prompting - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Ch.2 Generative Models - Foundations of Large Language Models
Related
Example of a Zero-Shot COT Prompt
Comparison of Few-Shot and Zero-Shot CoT Prompting
Alternative Phrases for Triggering Chain-of-Thought Reasoning
A user wants a large language model to solve a multi-step word problem. The model's initial attempts provide only a final, incorrect answer. The user's goal is to modify the prompt to encourage the model to generate a detailed, step-by-step thought process first, which should lead to a more accurate final answer. Crucially, the user does not want to include a complete, solved example of another problem in the prompt. Which of the following prompt modifications best achieves this specific goal?
To successfully prompt a language model to generate a step-by-step thought process for a new problem, one must always include a complete, solved example of a similar problem within the prompt.
Structure of a Zero-Shot CoT Prompt for an Arithmetic Task
Identifying a Zero-Shot Reasoning Prompt
Your team is rolling out an internal LLM assistant...
You’re building an internal LLM workflow to produc...
You’re building an internal LLM assistant to help ...
You’re leading an internal enablement team buildin...
Choosing and Justifying a Prompting Strategy Under Context and Quality Constraints
Designing a Prompting Workflow for a High-Stakes, Multi-Step Task
Diagnosing and Redesigning a Prompting Approach for a Decomposed Workflow
Stabilizing an LLM Workflow for Multi-Step Policy Compliance Decisions
Debugging a Multi-Step LLM Workflow for Contract Clause Risk Triage
Designing a Robust Prompting Workflow for Multi-Step Root-Cause Analysis with Limited Examples
Zero-Shot CoT Example with Jack's Apples
Learn After
A user wants a large language model to solve a simple arithmetic word problem by explaining its reasoning process. The user does not want to provide any pre-solved examples in the prompt. Which of the following prompts is best structured to achieve this?
Constructing a Reasoning-Eliciting Prompt
Diagnosing an Ineffective Prompt for Arithmetic Reasoning