Concept

Fixed-Size Window Memory as a Form of Local Attention

A simple and effective method for creating a fixed-size memory component, denoted as Mem\mathrm{Mem}, in attention mechanisms is to use a sliding window. This approach, which is a form of local attention, considers only a limited, constant number of the most recent key and value pairs. By restricting attention to this local neighborhood, the memory size is capped and prevented from growing with the input sequence length.

Image 0

0

1

Updated 2026-04-22

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences