Learn Before
Low Rank Self-Attention
The self-attention A has been observed to be low ranking, meaning that the rank of A is far lower than input length 𝑇. This implies that the low-rank property could be explicitly modeled with parameterization. Low rank self-attention is when the self-attention matrix is replaced by a low-rank approximation.
0
1
Tags
Data Science
Related
Sparse Attention
Query Prototyping and Memory Compression
Low Rank Self-Attention
Attention with Prior
Improved Multi-Head Attention Mechanism
Linear Attention
A research team is working to reduce the computational cost of the attention mechanism for processing extremely long documents. Their proposed solution involves modifying the attention calculation so that each query token only computes attention scores with a small, fixed subset of key tokens (e.g., neighboring tokens and a few globally important tokens) instead of all tokens in the sequence. Which category of attention improvement best describes this approach?
Match each attention improvement strategy with its core operational principle.
Optimizing Transformer Attention for Long Sequences
Evaluating Attention Optimization Strategies for Specific Applications