Learn Before
Language Model Performance Analysis
A language model was fine-tuned with a dataset consisting exclusively of instructions in two formats: 'Translate the following English text to French: [text]' and 'Summarize this article: [article]'. After training, the model performs perfectly on instructions matching these formats. However, when given the new prompt, 'Can you briefly explain the main points of this article and then provide a French translation of your explanation?', the model only provides a summary and does not attempt the translation. Based on this outcome, analyze the most likely reason for the model's partial failure. What fundamental challenge in training helpful models does this scenario illustrate?
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Language Model Performance Analysis
An AI development team fine-tunes two language models. Model A is trained on 100,000 examples of a single, narrow task: rephrasing sentences into five specific styles. Model B is trained on 10,000 examples covering a wide variety of tasks (e.g., summarization, translation, creative writing). When both models are tested on a completely new, unseen instruction like 'generate a grocery list for a three-course Italian meal,' which outcome is most likely?
Evaluating Training Data Strategies for Model Performance