Learn Before
Activity (Process)

Supervised Fine-Tuning (SFT)

Supervised Fine-Tuning (SFT) is a direct method for adapting pre-trained Large Language Models to follow instructions by training them on a dataset of annotated input-output pairs. In contrast to the pre-training objective of maximizing the probability of an entire sequence, SFT's goal is to maximize the conditional probability of generating the correct output given the input prefix. This process, formalized as Maximum Likelihood Estimation (MLE), teaches the model to produce the desired 'gold-standard' response for a given instruction.

Image 0

0

1

Updated 2026-04-30

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Ch.2 Generative Models - Foundations of Large Language Models

Related