True/False

Consider the listwise loss formula used for training on ranked preferences:

L=E[1N(N1)ya,ybYyayblogPr(yaybx)]\mathcal{L} = -\mathbb{E}\left[\frac{1}{N(N-1)}\sum_{\substack{\mathbf{y}_a, \mathbf{y}_b \in Y \\ \mathbf{y}_a\neq \mathbf{y}_b}} \log\Pr(\mathbf{y}_a \succ \mathbf{y}_b|\mathbf{x})\right]

True or False: If a model is completely uncertain about the preferences within a ranked list (i.e., it assigns Pr(yaybx)=0.5\Pr(\mathbf{y}_a \succ \mathbf{y}_b|\mathbf{x}) = 0.5 for all distinct pairs), the contribution of that specific list to the overall loss will be zero.

0

1

Updated 2025-10-08

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science