Short Answer

Explaining Positional Encoding Failure

A language model is trained exclusively on text segments with a maximum length of 1,024 tokens. When an analyst visualizes the model's positional signals for a 2,000-token input, they observe a structured, meaningful pattern for the first 1,024 positions, but a completely chaotic and noisy pattern for all subsequent positions. Based on this observation, explain the underlying mechanism that causes this specific pattern of failure.

0

1

Updated 2025-10-06

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science