Concept

Generalizable Positional Embeddings

To overcome the limitations of fixed-length training, an alternative approach is to develop generalizable positional embeddings. Suppose an embedding model is trained on sequences with a maximum length of mlm_l. If the model can generalize, it can be applied to handle much longer sequences of length mm (where mmlm \gg m_l) during inference. This capability allows the model to extrapolate and effectively deal with new positions outside the range observed in the training data.

0

1

Updated 2026-04-23

Contributors are:

Who are from:

Tags

Foundations of Large Language Models

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences