Learn Before
Deep Learning Optimizer Algorithms
When a neural network trains, it uses an algorithm to determine the optimal weights for the model, called an optimizer. There are several optimizer algorithms, such as:
- Gradient descent
- Mini-batch gradient descent
- Gradient descent with momentum
- RMSprop
- Adam optimization algorithm
- Nesterov momentum
- AdaGrad
0
2
Tags
Data Science
Related
Depth and Width for Neural Networks
Dropout
Neural Network Learning Rate
Epochs in Machine Learning
Activation Functions in Neural Networks
Deep Learning Optimizer Algorithms
Batch Normalization in Deep Learning
Deep Learning Weight Initialization
Hyperparameters Tuning Methods in Deep Learning
Difference between Model Parameter and Model Hyperparameter
Regularization Constant
Learn After
Mini-Batch Gradient Descent
Gradient Descent with Momentum
An overview of gradient descent optimization algorithms
Learning Rate Decay
Gradient Descent
AdaDelta (Deep Learning Optimization Algorithm)
Adam (Deep Learning Optimization Algorithm)
RMSprop (Deep Learning Optimization Algorithm)
AdaGrad (Deep Learning Optimization Algorithm)
Nesterov momentum (Deep Learning Optimization Algorithm)
Challenges with Deep Learning Optimizer Algorithms
Adam optimization algorithm
Difference between Adam and SGD