Multiple Choice

An LLM inference server is processing a batch of three long-running requests. In the middle of this process, after several computational steps have already been completed for the initial batch, a new, short request arrives. How would a system implementing continuous batching most likely handle this new request in the next computational step?

0

1

Updated 2025-10-05

Contributors are:

Who are from:

Tags

Ch.5 Inference - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science