Case Study

Debugging a Text Generation System

An engineer is debugging a text generation model that uses a search algorithm to build sentences. The model is producing very predictable and often repetitive outputs. For example, when prompted to complete 'The weather today is...', it consistently generates 'The weather today is nice. The weather today is nice.' Upon inspecting the generation process, the engineer notes that at each step, only the single most probable next word is ever considered to extend the current sequence.

Based on this observation, what specific aspect of the token selection process is likely causing this issue, and how should it be adjusted to encourage more diverse and potentially higher-quality outputs? Explain your reasoning.

0

1

Updated 2025-10-05

Contributors are:

Who are from:

Tags

Ch.5 Inference - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science