Learn Before
Concept

Sequential Token Processing in Recurrent Models

In long sequence modeling, recurrent models operate by reading one or a few tokens at a time. They use these inputs to update their internal recurrent state and then discard the inputs before the next tokens arrive. At any given step, the model generates its output based solely on the current recurrent state, rather than referring back to all previous states or past inputs.

0

1

Updated 2026-04-22

Contributors are:

Who are from:

Tags

Foundations of Large Language Models

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences