True/False

True or False: When training a reward model using the loss function L = E[(human_score - predicted_reward)^2], the primary objective is to ensure that for any two outputs, the one with the higher human score also receives a higher predicted reward from the model.

0

1

Updated 2025-10-08

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science