Learn Before
Concept

Batch Norm Makes Training Deep Networks Easier

Since we normalize the inputs in each layer, the changes in the previous layer won't have much influence on the next layer. This is because, it reduces the amount of shift in the distribution of inputs to each layer (covariate shifts).

0

1

Updated 2026-05-03

Tags

Data Science

D2L

Dive into Deep Learning @ D2L