Concept

Formulating the Loss Function for Policy Learning in RLHF

In the policy learning stage of Reinforcement Learning from Human Feedback (RLHF), after the LLM generates outputs for an input-only dataset, a loss function is formulated. This function is essential for quantifying the model's performance and guiding the update of its policy parameters.

Image 0

0

1

Updated 2026-05-02

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Related