Learn Before
KV Cache Structure Trade-offs
Analyze the two proposed modifications below for reducing the memory footprint of a model's Key-Value (KV) cache during text generation. For each option, describe how it alters the cache's multi-dimensional structure and discuss the likely trade-off for the model's ability to understand context.
0
1
Tags
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
An engineer modifies a large language model by doubling the number of attention heads per layer while simultaneously halving the dimensionality of each head's key/value vectors. Assuming all other parameters (like the number of layers and sequence length) remain constant, how does this architectural change affect the multi-dimensional structure of the model's key-value (KV) cache?
KV Cache Structure Trade-offs
Calculating KV Cache Size per Token