Learn Before
Example

Illustration of Pipeline Parallelism with Micro-batches

An illustration of pipeline parallelism typically demonstrates how computation is staggered across multiple workers (e.g., LL workers) to process multiple micro-batches. Let Bl,k\mathrm{B}_{l,k} denote the processing of the kk-th micro-batch by the ll-th worker. A pipeline is created where a subsequent worker begins processing a micro-batch immediately after the preceding worker has completed its step and passed it along. This staggered, overlapping execution allows multiple workers to be active concurrently on different micro-batches, which significantly maximizes hardware utilization and minimizes the idle time that occurs in simpler sequential approaches.

0

1

Updated 2026-04-21

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Computing Sciences

Foundations of Large Language Models Course

Related