Concept

Positional Encoding without Generalization

Positional encoding methods that lack generalization capabilities fail to generate meaningful values for sequence positions beyond the maximum length encountered during training. When visualized, the output for these longer sequences often appears chaotic or random, indicating the model's inability to understand positional relationships outside its training distribution.

0

1

Updated 2026-04-23

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences