Learn Before
Balancing General and Specific Knowledge in Model Training
A large language model first undergoes a pre-training phase on a massive, diverse dataset, followed by a supervised fine-tuning phase on a smaller, more specialized dataset. Analyze the distinct contribution of each training phase to the model's final abilities. In your analysis, discuss the primary challenge that arises when trying to add new, specific capabilities during the second phase without diminishing the broad knowledge gained in the first.
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
A development team starts with a large language model that has been pre-trained on a vast corpus of text from the internet, giving it a broad base of general knowledge. To make it a better customer service assistant, they then fine-tune it on a specific dataset of support chat logs. After this fine-tuning, they observe that while the model excels at customer service conversations, its performance on general trivia questions has noticeably degraded. What does this outcome most directly illustrate?
Chatbot Development Strategy
Balancing General and Specific Knowledge in Model Training