Learn Before
An engineer is optimizing a model for processing extremely long text sequences. To reduce the computational load, the model is designed so that each token primarily attends to a limited, local neighborhood of other tokens. The engineer observes that the model struggles to connect information from the end of a document back to key concepts introduced in the very first paragraph. Which of the following modifications best addresses this issue by providing a form of global context without sacrificing the overall computational efficiency?
0
1
Tags
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Performance Stabilization via Global Tokens
Trade-off of Fixed-Size Global Memory
An engineer is optimizing a model for processing extremely long text sequences. To reduce the computational load, the model is designed so that each token primarily attends to a limited, local neighborhood of other tokens. The engineer observes that the model struggles to connect information from the end of a document back to key concepts introduced in the very first paragraph. Which of the following modifications best addresses this issue by providing a form of global context without sacrificing the overall computational efficiency?
Analyzing Attention Mechanisms for Long Sequences
Evaluating a Hybrid Attention Strategy