Learn Before
Definition

Input Token Sequence in Language Models

In language modeling, an input sequence is mathematically defined as a sequence of tokens x0,x1,,xmx_0, x_1, \dots, x_m. The initial token, x0x_0, functions as a special start symbol that marks the beginning of the sequence. This symbol is frequently denoted as s\langle s \rangle or SOS\langle \mathrm{SOS} \rangle, while architectures such as BERT represent it using the [CLS][\mathrm{CLS}] token.

0

1

Updated 2026-04-18

Contributors are:

Who are from:

Tags

Ch.5 Inference - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Ch.2 Generative Models - Foundations of Large Language Models

Ch.1 Pre-training - Foundations of Large Language Models

Related