Concept

Fixed-Size Memory for Constant Attention Cost

If the memory component Mem\mathrm{Mem} used in the attention operation is defined as a fixed-size variable, the computational cost of performing the attention function Att(qi,Mem)\mathrm{Att}(\mathbf{q}_i, \mathrm{Mem}) will be fixed. By representing keys and values using this fixed-size memory model, the cost remains constant regardless of the sequence length. This foundational concept opens up several alternative ways to design the memory Mem\mathrm{Mem}.

0

1

Updated 2026-04-22

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences