Multiple Choice

In the context of a reward-weighted probability distribution, defined as π(yx)=πθref(yx)exp(1βr(x,y))Z(x)\pi^{*}(\mathbf{y}|\mathbf{x}) = \frac{\pi_{\theta_{\text{ref}}}(\mathbf{y}|\mathbf{x}) \exp \left(\frac{1}{\beta}r(\mathbf{x}, \mathbf{y})\right)}{Z(\mathbf{x})}, consider a scenario where a specific output, yA\mathbf{y}_A, receives a very high reward, r(x,yA)r(\mathbf{x}, \mathbf{y}_A). However, the reference distribution assigns a probability to this output that is extremely close to zero, i.e., πθref(yAx)0\pi_{\theta_{\text{ref}}}(\mathbf{y}_A|\mathbf{x}) \approx 0. What will be the approximate probability of yA\mathbf{y}_A in the final distribution, π(yAx)\pi^{*}(\mathbf{y}_A|\mathbf{x})?

0

1

Updated 2025-10-08

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science