Learn Before
Definition

Set of Distributed Data Batches in Data Parallelism

In data parallelism, a minibatch of training sample, Dmini\mathcal{D}_{\mathrm{mini}}, is divided into NN smaller batches, which can be denoted by D1,...,DN\mathcal{D}^{1},...,\mathcal{D}^{N}. After the division, these smaller batches are distributed to NN separate workers, each receiving one corresponding batch, allowing them to work at the same time.

Image 0

0

1

Updated 2026-04-21

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Related