Multiple Choice

A machine learning engineer is training a small 'student' model to mimic a large 'teacher' model. The training process aims to minimize the Kullback-Leibler (KL) divergence between the teacher's output probability distribution (P_teacher) and the student's (P_student), formulated as: Loss = KL(P_teacher || P_student). Based on the properties of this specific formulation, what is the primary effect of minimizing this loss on the student model's behavior?

0

1

Updated 2025-09-28

Contributors are:

Who are from:

Tags

Ch.3 Prompting - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science