Learn Before
A development team has successfully used a distributed computing strategy to spread a large model's computational work across multiple devices during its initial training phase. They now plan to use this exact same distributed setup to run the model for a live, user-facing application. Which statement best analyzes the viability of this plan?
0
1
Tags
Ch.5 Inference - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Mixture-of-Experts (MoE) for Efficient Inference
Challenges in Applying Parallelization to LLM Inference
Applicability of Pre-training Parallelism Strategies to LLM Inference
Complexity of LLM Serving Systems
A development team has successfully used a distributed computing strategy to spread a large model's computational work across multiple devices during its initial training phase. They now plan to use this exact same distributed setup to run the model for a live, user-facing application. Which statement best analyzes the viability of this plan?
Scaling an LLM-Powered Service
Match each parallelization strategy with the description of how it distributes computational work across multiple devices.