Learn Before
Short Answer

Positional Information in Vector Transformations

A language model processes the sentence 'The quick brown fox jumps'. The token 'brown' is at position 3, and the token 'jumps' is at position 5. Later, it processes a much longer text where the same two tokens appear as '...the lazy brown dog jumps...', with 'brown' now at position 42 and 'jumps' at position 44. If the model encodes position by applying a rotational transformation to each token's vector, what fundamental aspect of the relationship between the vectors for 'brown' and 'jumps' will remain consistent across both contexts? Explain why this consistency is beneficial for the model.

0

1

Updated 2025-10-02

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Application in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science

Related