Concept
Batch Norm Makes Training Deep Networks Easier
Since we normalize the inputs in each layer, the changes in the previous layer won't have much influence on the next layer. This is because, it reduces the amount of shift in the distribution of inputs to each layer (covariate shifts).
0
1
Updated 2026-05-03
Contributors are:
Who are from:
Tags
Data Science
D2L
Dive into Deep Learning @ D2L
Related
Why do we normalize the inputs X in Deep Learning?
Visualization of How Batch Norm Helps Speed up Deep Learning
Batch Norm Makes Training Deep Networks Easier
Batch Norm as Regularization in Deep Learning
Reduction of Covariate Shift via Layer Normalization
Batch Norm Makes Training Deep Networks Easier
Medical Diagnostics Example of Covariate Shift