Learn Before
Analyzing Context Limitations in a Recurrent Model
Based on the model's described memory architecture, what is the most likely reason for this failure to connect the character's action to their motivation?
0
1
Tags
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
A language model processes a long document by dividing it into 10 equal, non-overlapping segments. To maintain context, the model's attention mechanism at any point can access information from the segment it is currently processing as well as the single segment that came immediately before it. If the model is currently processing Segment 6, which segments' information is available to its attention mechanism?
Analyzing Context Limitations in a Recurrent Model
Analyzing Memory Trade-offs in Segment-Level Recurrence
Compressive Transformer Memory Architecture