Activity (Process)

Greedy Decoding in Language Models

In language model inference, a common method for generating text is to select the token with the maximum probability from the predicted distribution at each position. This strategy is applied sequentially, where the model's output at each step is determined by the single most likely next token given the preceding sequence.

0

1

Updated 2025-10-10

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences