Concept

Fine-Tuning as a Mechanism for Activating Pre-Trained Knowledge

The pre-training and fine-tuning paradigm operates on the principle that LLMs acquire latent abilities for instruction comprehension and response generation during pre-training. However, these learned instruction-response mappings may not have a high probability of being generated during inference. Fine-tuning serves as a mechanism to activate these dormant capabilities by slightly adjusting the model's parameters using a small set of supervised data, which increases the likelihood of generating the desired responses to instructions.

0

1

Updated 2026-05-02

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Ch.4 Alignment - Foundations of Large Language Models

Related