Learn Before
Multiple Choice

A language model is tasked with answering a question by identifying the correct text span from a given context. The model works by calculating a probability for each token being the 'start' of the answer and a separate probability for each token being the 'end' of the answer. Consider the following scenario:

Context: 'The first modern Olympic Games were held in Athens, Greece, in 1896. The International Olympic Committee (IOC) was founded in 1894 by Pierre de Coubertin.' Question: 'When was the IOC established?'

The model produces the following highest probabilities:

  • Highest Probability Start Token: '1896' (Probability: 0.85)
  • Highest Probability End Token: '1894' (Probability: 0.91)

Based on this output, what is the most fundamental reason the model failed to produce a valid answer span?

0

1

Updated 2025-09-28

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science