Multiple Choice

A language model combines a base probability distribution, P_lm, with a retrieval-based distribution, P_knn, to predict the next token. The final probability is calculated by blending these two distributions using an interpolation coefficient, λ = 0.6. Given the distributions below for a small vocabulary, which token will the model select as its final output?

  • P_lm: {'wordA': 0.5, 'wordB': 0.4, 'wordC': 0.1}
  • P_knn: {'wordA': 0.2, 'wordB': 0.7, 'wordC': 0.1}

0

1

Updated 2025-09-28

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Application in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science