Case Study

Debugging a Language Model's Output Distribution

Considering the final steps of next-token probability calculation (transformation of the final hidden state to logits, followed by the application of a function to get probabilities), what is the most plausible cause of this behavior? Explain your reasoning.

0

1

Updated 2025-10-09

Contributors are:

Who are from:

Tags

Ch.5 Inference - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science