Concept

Establishing the Initial Policy in RLHF

The starting point for Reinforcement Learning from Human Feedback (RLHF) is an initial policy, which is an LLM that has already undergone pre-training and instruction fine-tuning. This model is considered the version that would be deployed to interact with users and respond to their requests, forming the baseline for further alignment.

0

1

Updated 2026-04-20

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences