Limitations of Minimal Data Fine-Tuning
A research team successfully fine-tuned a large pre-trained model to follow instructions for summarizing news articles using only 100 high-quality examples. However, when they later asked the same model to write a short poem, it produced a summary of the word 'poem' instead. Based on the principles of data-efficient fine-tuning, explain the most probable reason for this specific failure.
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Evaluating a Data-Efficient Fine-Tuning Strategy
A small startup has access to a large, pre-trained language model. Their goal is to make the model generate social media posts that perfectly match their company's unique and witty brand voice. Given their limited budget and time, which of the following strategies represents the most data-efficient approach to achieve this specific instruction-following behavior?
Limitations of Minimal Data Fine-Tuning