Learn Before
Multiple Choice

An engineer designs a sparse attention mechanism where, for any given token at position i, the model is only allowed to attend to the tokens within a fixed-size window around it (e.g., from position i-k to i+k). This rule is applied uniformly across the entire sequence, irrespective of the specific words involved. Which statement best analyzes the core principle of this design?

0

1

Updated 2025-09-26

Contributors are:

Who are from:

Tags

Data Science

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science