Learn Before
Attention-level improvements of Transformers
Improvements to the attention module can be categorized into these directions:
- Sparse Attention
- Linearized Attention
- Prototype and Memory Compression
- Low-rank Self-Attention
- Attention with Prior
- Improved Multi-Head Attention
0
1
Tags
Data Science
Foundations of Large Language Models Course
Computing Sciences
Learn After
Sparse Attention
Query Prototyping and Memory Compression
Low Rank Self-Attention
Attention with Prior
Improved Multi-Head Attention Mechanism
Linear Attention
A research team is working to reduce the computational cost of the attention mechanism for processing extremely long documents. Their proposed solution involves modifying the attention calculation so that each query token only computes attention scores with a small, fixed subset of key tokens (e.g., neighboring tokens and a few globally important tokens) instead of all tokens in the sequence. Which category of attention improvement best describes this approach?
Match each attention improvement strategy with its core operational principle.
Optimizing Transformer Attention for Long Sequences
Evaluating Attention Optimization Strategies for Specific Applications