Learn Before
Teacher-Student Model Architecture in Knowledge Distillation
In a knowledge distillation framework, a larger and more powerful 'teacher' model is used to train a 'student' model that is designed to be smaller and more efficient. The teacher model processes a full-context user input to generate its output probability, denoted as . In contrast, the student model processes a simplified context input to produce its own output, . The training objective is to transfer knowledge from the stronger teacher to the compact student by minimizing a loss function that measures the difference between their respective outputs.
0
1
Tags
Ch.3 Prompting - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Ch.4 Alignment - Foundations of Large Language Models
Learn After
Distillation Loss for Response-Based Knowledge
Objective Function for Student Model Training via Knowledge Distillation
Definition of Teacher's Probability Distribution (Pt) in Knowledge Distillation
Definition of Student's Probability Distribution (P_theta^s)
General Loss Function for Knowledge Distillation
Optimizing a Language Model for Mobile Deployment
Definition of Student's Probability Distribution ()
A research lab has developed a very large and complex language model that achieves state-of-the-art performance on a translation task. However, due to its size, the model is too slow and expensive to deploy for a real-time translation mobile app. To address this, the team uses the large model's predictions on a set of sentences to train a new, much smaller and faster model. What is the primary strategic advantage of this two-model approach?
A development team is using a knowledge distillation framework to create a compact, efficient language model (the 'student') from a much larger, high-performance model (the 'teacher'). The goal is to deploy the student model on devices with limited computational resources. Which statement best analyzes the typical relationship between the inputs processed by the teacher and student models during this process?