True/False

In a causal attention mechanism that incorporates relative positional information, consider the calculation of attention for an output at position i. If the dot product of the query vector from position i with the key vector from position j is identical to its dot product with the key vector from position k (where j ≠ k, and both j, k < i), then the final attention weights assigned to positions j and k will also be identical.

0

1

Updated 2025-10-08

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science