Definition

Special Tokens in Language Models

Language models utilize special tokens that do not represent words from the vocabulary but instead provide structural or control information. For instance, the ⟨SOS⟩ (Start of Sequence) token marks the beginning of an input, while the ⟨pad⟩ token is used to equalize sequence lengths within a batch for efficient processing. These tokens are essential for managing the input and output data streams of the model.

0

1

Updated 2025-10-08

Contributors are:

Who are from:

Tags

Ch.5 Inference - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences