Concept

Weight Decay

Weight decay, commonly known as 2\ell_2 regularization, is a widely used technique for regularizing parametric machine learning models. Instead of directly manipulating the number of parameters, weight decay operates by restricting the values that the parameters can take. The technique is motivated by the intuition that the simplest function is f=0f = 0, and the complexity of a linear function, such as f(x)=wopxf(\mathbf{x}) = \mathbf{w}^ op \mathbf{x}, can be measured by the distance of its parameters from zero.

0

2

Updated 2026-05-03

Tags

Data Science

D2L

Dive into Deep Learning @ D2L