Learn Before
A development team is fine-tuning a large language model for a specific task using a dataset of inputs and corresponding correct outputs. During a training iteration, the model produces an output that is very different from the correct target output. What is the immediate, primary function of this discrepancy within the training process?
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Application in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Classification of LLM Adaptation Methods
RLHF Policy Optimization as Loss Minimization
A development team is fine-tuning a large language model for a specific task using a dataset of inputs and corresponding correct outputs. During a training iteration, the model produces an output that is very different from the correct target output. What is the immediate, primary function of this discrepancy within the training process?
Direct Supervision via Knowledge Distillation Loss in Weak-to-Strong Generalization
A large language model is undergoing a single step of fine-tuning on a new dataset. Arrange the following events in the correct chronological order to represent this process.
Data Selection and Filtering using Small Models
Diagnosing a Stagnant Fine-Tuning Process