Concept

Risk of Overfitting and Catastrophic Forgetting in SFT

Supervised Fine-Tuning presents a paradoxical challenge: while it aims to incorporate new knowledge, the process of updating a pre-trained model can lead to the loss of its original knowledge. This issue, known as catastrophic forgetting, is particularly pronounced when an LLM is fine-tuned extensively on a large SFT dataset. Such intensive training can cause the model to overfit the new data, which not only harms its ability to generalize to unseen examples but can also erase foundational knowledge acquired during pre-training.

0

1

Updated 2026-05-01

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences