Analyzing Task Ambiguity Resolution
Based on the provided scenario, analyze why the addition of the first two examples ('London -> UK', 'Rome -> Italy') leads the model to produce the desired output. Explain this phenomenon in terms of how the model's general knowledge is being directed.
0
1
Tags
Ch.3 Prompting - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Analyzing Task Ambiguity Resolution
A user wants a Large Language Model to perform a specific task: extract only the primary company name from a news headline. The model's broad pre-training means it could mistakenly extract names of people, products, or other organizations.
The final headline to be processed is: 'Tech giant InnovateCorp announces a new partnership with Global Logistics.'
Analyze the two sets of in-context examples below. Which set provides a better guiding mechanism for the model to correctly identify 'InnovateCorp' as the desired output, and what is the most accurate reason?
Set A:
- Headline: 'QuantumLeap Inc. reveals breakthrough in computing.' -> QuantumLeap Inc.
- Headline: 'Shares of AutoDrive Solutions soar after earnings report.' -> AutoDrive Solutions
Set B:
- Headline: 'CEO John Smith discusses future of AI.' -> John Smith
- Headline: 'New smartphone 'Photon' to be released next month.' -> Photon
Guiding LLM Summarization