Concept

Next Token Prediction Task

When applying a trained language model, a common and fundamental task is next token prediction, which involves finding the most likely token given its sequence of previous context tokens. At each step, the model computes a probability distribution over the entire vocabulary conditioned on the preceding context. This token prediction task sequentially utilizes these computed probability distributions to select the most probable next token and continue the sequence.

Image 0

0

1

Updated 2026-05-02

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences