Multiple Choice

A language model is generating text one token at a time by always selecting the single most probable next token. It has already produced the sequence 'The sun is shining'. For the very next step, the model calculates the following conditional probabilities for the next token:

  • P(brightly | 'The sun is shining') = 0.55
  • P(today | 'The sun is shining') = 0.25
  • P(and | 'The sun is shining') = 0.15
  • P(down | 'The sun is shining') = 0.05

Based on this method of construction, what will the updated sequence be after this step?

0

1

Updated 2025-09-28

Contributors are:

Who are from:

Tags

Ch.5 Inference - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Application in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science