Short Answer

The Reward Model's Functional Shift

A reward model is initially trained to predict which of two responses a human would prefer. Later, this same model is used to assign a single numerical score to individual responses to guide the fine-tuning of a large language model. Explain the fundamental difference between the model's task during its training phase and its application phase, and analyze why this shift is crucial for the fine-tuning process.

0

1

Updated 2025-10-06

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Ch.4 Alignment - Foundations of Large Language Models

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science