Learn Before
Diagnosing a Failing Data Labeling Pipeline
A data scientist is using a Large Language Model to automatically label thousands of customer reviews for a sentiment analysis project. Their goal is to create a dataset where each review is tagged with one of three exact labels: 'Positive', 'Negative', or 'Neutral'. However, their automated script keeps failing. They discover the model is producing a wide variety of responses, such as 'The sentiment of this text is clearly positive.', 'I'd say this is a negative review.', or simply 'Neutral'. Based on this information, analyze the likely cause of the script's failure and describe the specific modification you would make to the prompt to ensure the model produces consistent, machine-readable output.
0
1
Tags
Ch.3 Prompting - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Example of a Prompt with a Post-Input Instruction for Polarity Classification
A developer is building an automated system to sort customer support tickets. The system requires a language model to categorize each ticket's sentiment so it can be routed correctly. For the system to work, the model's output must be exactly one of the following words: 'Positive', 'Negative', or 'Neutral'. Which prompt most effectively uses a direct, imperative command to achieve this specific, constrained output?
Crafting a Constraining Command for an LLM
Diagnosing a Failing Data Labeling Pipeline