A developer observes that a standard Transformer-based language model takes approximately 2 seconds to process a text sequence of 500 tokens. Based on the computational properties of the model's core mechanism, what is the most likely processing time if the input sequence length is doubled to 1000 tokens?
0
1
Tags
Ch.5 Inference - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Application in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Language Model Performance Analysis
A developer observes that a standard Transformer-based language model takes approximately 2 seconds to process a text sequence of 500 tokens. Based on the computational properties of the model's core mechanism, what is the most likely processing time if the input sequence length is doubled to 1000 tokens?
Model Selection for Long-Document Summarization