Concept

Function to Measure Differences Between Models

In architectures that utilize both small and large models, a mathematical function is often defined to quantify the difference between their outputs. For example, in knowledge distillation, the Kullback-Leibler (KL) divergence can be used to measure the dissimilarity between the probability distributions generated by a large 'teacher' model and a smaller 'student' model for a given input.

Image 0

0

1

Updated 2026-05-02

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences