Activity (Process)

Weak-to-Strong Generalization via Fine-Tuning on Weak Model Data

One approach to weak-to-strong generalization involves a two-stage process. First, a dataset is curated using a small, weak model. This can be done either by having the weak model generate labels for a set of inputs or by using it to select high-quality examples from a larger, pre-existing dataset. In the second stage, a large, strong model is fine-tuned on this curated dataset. The training objective is to minimize a loss function, such as a Knowledge Distillation (KD) loss, which measures the discrepancy between the strong model's outputs and the labels provided by the weak model in the dataset.

Image 0

0

1

Updated 2026-05-02

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Related
Learn After