Log-Probability of a Ranked Sequence
The log-probability of an ordered list or ranked sequence of preferences, , given an input , can be defined as the sum of the conditional log-probabilities at each stage of selection. For an ordered list given by , the overall log-probability is expressed as: where represents the subset of the list of outputs that remain unselected at the -th stage, i.e., .

0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
Applying the Plackett-Luce Model to RLHF Reward Modeling
Log-Probability of a Ranked Sequence
An AI team is using a probabilistic model to rank three generated summaries (A, B, C). The model assigns a positive 'strength' score to each summary. The probability of a summary being chosen as best from a given set of options is its strength score divided by the sum of the strength scores of all summaries in that set. This selection process is repeated to form a full ranking. Given the scores below, which statement is correct?
- Summary A Strength: 6.0
- Summary B Strength: 3.0
- Summary C Strength: 1.0
An AI system uses a probabilistic model to rank three generated text snippets: Snippet A, Snippet B, and Snippet C. The model assigns a positive 'worth' score to each snippet (A=9, B=6, C=3). The probability of a specific ranking is found by sequentially calculating the probability of choosing the best snippet from the remaining set of options. Arrange the following steps in the correct order to calculate the probability of the ranking A > B > C.
Calculating Ranking Probability
Log-Probability of a Ranked Sequence
Log-Likelihood Objective for Language Model Training
A language model is generating a sequence of tokens. It has computed the following conditional log-probabilities for a three-token sequence, where each token's probability is dependent on the ones that came before it:
- Log-probability of the first token: -1.8
- Log-probability of the second token, given the first: -2.5
- Log-probability of the third token, given the first two: -1.2
Based on these values, what is the total log-likelihood of this entire three-token sequence?
Evaluating Sentence Plausibility
A language model has calculated the total log-likelihood for the sequence of tokens: ["The", "quick", "brown", "fox"]. The calculation involves summing the conditional log-probabilities of each token given the preceding ones. If the third token is changed from "brown" to "lazy", creating the new sequence ["The", "quick", "lazy", "fox"], which set of conditional log-probabilities must be re-calculated to find the new total log-likelihood?
Applying Log-Likelihood Calculation to a Training Dataset
Learn After
Set of Remaining Items in a Ranked Sequence
Plackett-Luce Loss Function
A model is designed to rank a set of three documents {Doc A, Doc B, Doc C} for a given user query. To calculate the log-probability of the specific ranked sequence 'Doc A > Doc B > Doc C', a developer proposes calculating the total log-probability as the sum of the log-probabilities of each document being chosen first from the full set of three documents. Why is this approach fundamentally flawed for modeling a sequential ranking process?
Ranked Sequence Log-Probability Calculation
Calculating Log-Probability for a Ranked List