A research team fine-tunes a very capable, state-of-the-art model using labels generated by a much older, less powerful model. The team's goal is for the capable model to learn the underlying task better than its less powerful supervisor. However, after training, the capable model's performance on a standard evaluation set is worse than before the fine-tuning. What is the most likely reason for this failure?
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Computing Sciences
Foundations of Large Language Models Course
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
A research team fine-tunes a very capable, state-of-the-art model using labels generated by a much older, less powerful model. The team's goal is for the capable model to learn the underlying task better than its less powerful supervisor. However, after training, the capable model's performance on a standard evaluation set is worse than before the fine-tuning. What is the most likely reason for this failure?
Predicting Outcomes of Cross-Model Supervision
Surpassing the Supervisor