Example

Visual Diagram of Weak-to-Strong Generalization via Data Selection

This diagram illustrates a two-stage method for weak-to-strong generalization. In the first stage, a small, weaker model performs 'Data Selection' on an initial dataset to create a curated, higher-quality subset. In the second stage, a large, stronger model is fine-tuned on this selected data. The training loop involves the large model processing an input 'x' to produce an output, which is then compared against the corresponding label 'y' from the curated dataset. The discrepancy is used to compute a loss, often a Knowledge Distillation (KD) loss, which guides the training of the large model.

Image 0

0

1

Updated 2025-10-09

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Related