Long-Context LLMs
Long-context LLMs are a specialized category of Large Language Models capable of handling and processing extensive textual contexts. For instance, these models can read a source code file containing tens of thousands of lines to effectively outline the program's overall functionality.
0
1
Tags
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Ch.3 Prompting - Foundations of Large Language Models
Related
Classification of Long Sequence Modeling Problems
Increased Research Interest in Long-Context LLMs
Long-Context LLMs
Research Directions for Adapting Transformers to Long Contexts
Sparse Attention
Challenges in Training and Deploying High-Capacity Models
Challenge of Streaming Context for LLMs
Key Issues in Long-Context Language Modeling Methods
Challenge of Training New Architectures for Long-Context LLMs
Key Techniques for Long-Input Adaptation in LLMs
RoPE Scaling Transformation Equivalence
Architectural Prioritization for a Long-Context LLM
A development team is attempting to use a standard Transformer-based LLM for real-time analysis of continuous data streams, where the input sequence can grow to hundreds of thousands of tokens. They encounter two main problems: the time it takes to process each new token increases dramatically as the sequence gets longer, and the system frequently runs out of memory. Which statement correctly analyzes the architectural sources of these two distinct problems?
Differentiating Bottlenecks in Long-Sequence LLMs
Learn After
General Applicability of Long-Context Methods
Context Scaling for LLM Performance Improvement
Model Selection for Large-Scale Document Summarization
A development team is tasked with creating a system that can analyze and answer questions about lengthy legal documents, some of which are over 100,000 words long. When selecting a foundational language model for this task, what is the most critical architectural characteristic they should prioritize to ensure the system can effectively process the entirety of these documents at once?
Evaluating System Architectures for Long-Document Q&A
Infinite Context Encoding in LLMs
Continuous-Space Attention for Infinite Context