Learn Before
Optimizing Inference Performance
Based on the provided scenario, identify the most likely reason for the system's slow performance and propose a specific adjustment to the generation process to address it. What is the primary trade-off associated with your proposed adjustment?
0
1
Tags
Ch.5 Inference - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Application in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Diminishing Returns in Search Scaling
Optimizing Inference Performance
An engineer modifies a language model's inference procedure to evaluate a significantly larger number of potential output sequences at each generation step, aiming to enhance the final output quality. What is the most direct and unavoidable trade-off associated with this modification?
Resource Consumption in Text Generation