Learn Before
Concept
Common Learning Rate Decay Implementation
, where is the learning rate in the current epoch, is the initial learning rate, is the current epoch and is the decay rate selected. The decay rate is a tunable hyperparameter.
Initializing and , we graph an example with on x-axis and on y-axis. In the graph we observe the decay of learning rate.
0
2
Updated 2020-11-16
Contributors are:
Who are from:
Tags
Data Science
Related
Example Using Mini-Batch Gradient Descent (Learning Rate Decay)
Common Learning Rate Decay Implementation
Other Learning Rate Decay Implementations
Manual Implementation Learning Rate Decay
Learning Rate
Which of these statements about mini-batch gradient descent do you agree with?
Mini-Batch Gradient Descent Algorithm
Common Learning Rate Decay Implementation