Learn Before
Definition
Q/KV Heads Notation in LLM Architectures
When comparing the architectures of different Large Language Models (LLMs), the number of attention heads is often expressed in an format. In this specific notation, indicates the number of attention heads used for queries, while indicates the number of heads used for both keys and values.
0
1
Updated 2026-04-19
Tags
Foundations of Large Language Models
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences