Concept

Combined Loss Objective in Weak-to-Strong Training

When fine-tuning a large model with supervision from a weaker model, the training objective can be a composite loss function. This function often combines a Knowledge Distillation (KD) loss, which encourages the large model to imitate the weak model's outputs, with a standard Language Model (LM) loss. The LM loss is calculated against ground-truth labels, allowing the large model to learn from both the weak supervisor's generalized knowledge and high-quality annotated data simultaneously.

Image 0

0

1

Updated 2025-10-10

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences