Case Study

Reward Model Loss Calculation

An engineer is training two reward models, Model A and Model B, using a segment-based rating loss function. For a specific text segment, the human-provided target score is 4.0. Model A predicts a score of 3.9 for this segment, while Model B predicts a score of 3.0. Based on the principles of the rating loss function, which model will receive a stronger corrective signal (i.e., a larger loss value to minimize) for this specific segment, and why?

0

1

Updated 2025-09-29

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science