Example of Persistent General-Purpose Behavior: Math Fine-Tuning
An illustrative example of an LLM's persistent generalist nature is observed when a model is fine-tuned on mathematics data. Despite this specialized training, the model may not exclusively produce math-related outputs. If prompted to write poetry, for instance, it can still perform this general task, demonstrating that its core instruction-following ability remains broad.
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Computing Sciences
Foundations of Large Language Models Course
Related
Example of Persistent General-Purpose Behavior: Math Fine-Tuning
Using Diverse Data to Steer LLM Specialization
A development team adapts a large, pre-existing language model to function as a specialized chatbot for a legal information service. The adaptation process uses a dataset consisting solely of legal questions and their corresponding factual answers. After deployment, the team finds that the chatbot accurately answers legal queries but also responds correctly when users ask it to write poems or summarize news articles. Which statement provides the most accurate explanation for the chatbot's behavior?
Diagnosing Unexpected Model Behavior
Explaining Unintended Model Capabilities
Learn After
Unexpected Capabilities of a Fine-Tuned Chatbot
A development team fine-tunes a large language model on a massive dataset composed exclusively of historical texts from the 18th century, intending to create an expert assistant for historians. During a demonstration, an audience member asks the model to write a simple computer program to sort a list of numbers. The model successfully generates the correct code. Which of the following principles does this outcome best demonstrate?
Predicting LLM Behavior Post-Specialization