Comparison

Comparison of SFT and Pre-training Datasets

The dataset used for Supervised Fine-Tuning (SFT) differs significantly from the one used for pre-training. While the SFT dataset is considerably smaller in volume, its content is highly specialized and curated for the specific tasks the model is being adapted for.

0

1

Updated 2025-10-10

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Related