Calculating Sequence Log-Likelihood
A language model is being trained with the objective of maximizing the log-likelihood of sequences. For the specific sequence 'The cat sat', the model computes the following conditional log-probabilities for the actual next token at each position (assuming a fixed start-of-sequence token):
log Pr('The' | start_token) = -1.5log Pr('cat' | start_token, 'The') = -0.8log Pr('sat' | start_token, 'The', 'cat') = -1.2
Calculate the total log-likelihood for this entire sequence, which represents the value the model aims to maximize for this training example.
0
1
Tags
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Ch.5 Inference - Foundations of Large Language Models
Application in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Calculating Sequence Log-Likelihood
A language model is being trained on the sentence '
The cat sat'. The model calculates the following conditional log-probabilities at each step, where ' ' is a fixed start-of-sequence token: log P('The' | '<BOS>') = -1.5log P('cat' | '<BOS>', 'The') = -0.9log P('sat' | '<BOS>', 'The', 'cat') = -1.2
Based on the standard training objective for this single sequence, what is the total log-likelihood value that the model aims to maximize?
Model Output Evaluation
Youâre reviewing an internal evaluation script tha...
Your team is building an internal tool that ranks ...
Youâre reviewing an internal LLM evaluation pipeli...
Reconciling Training Log-Likelihood with Inference-Time Sequence Selection
Explaining a Counterintuitive Decoding Outcome Using Softmax, Next-Token Conditionals, and Sequence Log-Probability
Diagnosing a âHigh-Confidence Wrong Tokenâ Bug in Autoregressive Scoring
Investigating a Production Scoring Bug: Softmax Normalization vs. Autoregressive Sequence Log-Probability
Design a Correct Sequence-Scoring Function for Autoregressive LLM Outputs
Root-Cause Analysis: Why a âMore Likelyâ Token-by-Token Completion Loses on Total Sequence Score
Auditing a Candidate Completion Using Softmax Next-Token Probabilities and Autoregressive Log-Probability