Activity (Process)

Model Initialization Strategy in RLHF

The first step in the RLHF training process is to initialize the necessary models from existing ones. Typically, the reward and value models are initialized using a pre-trained Large Language Model. In contrast, the reference and target (policy) models are initialized from a model that has already undergone instruction fine-tuning. After this initialization, the reference model's parameters are fixed and are not updated further during the subsequent training stages.

Image 0

0

1

Updated 2026-05-02

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Ch.4 Alignment - Foundations of Large Language Models