Parameter Initialization and Moderate Adjustment in Fine-Tuning
In practical applications of fine-tuning, the model's parameters, denoted as , are initialized with the values from the pre-trained model. The subsequent adjustments to these parameters are intentionally moderate to ensure that the fine-tuned model retains the core knowledge from pre-training and does not deviate excessively from its original capabilities.
0
1
Tags
Ch.3 Prompting - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
Parameter Initialization and Moderate Adjustment in Fine-Tuning
Probabilistic Objective of Supervised Fine-Tuning
Comparison of Training Objectives: Instruction Fine-Tuning vs. Pre-training
A machine learning team has successfully developed a large language model by training it on a massive, general-purpose text corpus. They now want to make the model better at following specific user commands. To do this, they have created a new, high-quality dataset that is much smaller than the original corpus and consists of example commands paired with ideal responses. Based on the standard procedures for adapting such models, which statement best describes the relationship between the initial training phase and this new adaptation phase?
Training Strategy for a Specialized Chatbot
The training methodology for instruction fine-tuning must be fundamentally different from the methodology used for pre-training, primarily because the dataset used for fine-tuning is substantially smaller.
Objective of Instruction Fine-Tuning
Learn After
A machine learning engineer is adapting a large, pre-trained language model for a highly specialized task using a small dataset. They choose an aggressive training strategy that results in substantial changes to the model's parameters from their initial, pre-trained values. Which of the following outcomes represents the most significant risk of this approach?
Diagnosing Fine-Tuning Performance Degradation
Critiquing a Fine-Tuning Strategy