Analyzing Hardware Utilization in Batched Inference
Based on the provided system log, identify the specific iteration where a key inefficiency occurs for Sequence B and explain why this inefficiency is a direct consequence of the batching strategy described.
0
1
Tags
Ch.5 Inference - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
A language model processes a batch containing two sequences: Sequence A with a long prompt and Sequence B with a short prompt. The system is configured to complete the entire prompt-processing (prefill) phase for all sequences in the batch before starting the parallel token-generation (decode) phase for the entire batch. Which statement best analyzes the primary source of computational inefficiency in this scenario?
Analyzing Hardware Utilization in Batched Inference
Explaining Inefficiency in Batched Processing