Learn Before
Concept
Batch Size and Stability
Batch size controls the accuracy of the estimate of the error gradient when training neural networks. The model tends to stabilize more towards the end of the run. Smaller batch sizes are used for two main reasons: 1) Smaller batch sizes are noisy, offering a regularizing effect and lower generalization error. Smaller batch sizes make it easier to fit one batch worth of training data in memory (e.g. when using a GPU). The batch size is often set at some small values such as 32, and is not tuned by the practitioner.
1
3
Updated 2021-10-30
Tags
Data Science