Multiple Choice

A language model predicts the probabilities for the next word in a sequence. The top four candidates are: 'happy' (0.4), 'sad' (0.2), 'angry' (0.1), and 'joyful' (0.05). A decoding method is applied that restricts the possible choices to only the top three candidates ('happy', 'sad', 'angry'). After the probabilities for this smaller set are rescaled to form a new, valid probability distribution, what is the new probability for the word 'sad'?

0

1

Updated 2025-09-28

Contributors are:

Who are from:

Tags

Ch.5 Inference - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Application in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science