Learn Before
Concept
Attention Mechanism
An attention mechanism acts as a learnable pointer structure that allows a model to dynamically focus on specific parts of a sequence. This solves the challenge of increasing a system's memory and complexity without adding learnable parameters, and bypasses the bottleneck of compressing an entire sequence into a fixed-dimensional representation.
0
1
Updated 2026-05-01
Tags
D2L
Dive into Deep Learning @ D2L