Multiple Choice

A language model alignment method re-expresses the probability of a preferred response (y_a) over a dispreferred response (y_b) for a given prompt (x) as follows:

Pr(y_a ≻ y_b | x) = Sigmoid( β log( π_θ(y_a|x) / π_ref(y_a|x) ) - β log( π_θ(y_b|x) / π_ref(y_b|x) ) )

Where π_θ is the policy being trained and π_ref is a fixed reference policy. Based on this mathematical formulation, what is the primary reason this method can be trained without an explicit, separately-trained reward model?

0

1

Updated 2025-09-26

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science

Related