Gradient Descent
Gradient descent is a fundamental optimization algorithm that leverages gradients to minimize a model's loss function. Because the gradient of a function points in the direction of steepest ascent, moving the model's parameters in the opposite direction iteratively lowers the loss. Each step of such gradient-based optimization algorithms requires calculating the exact gradient of the loss with respect to the parameters.

0
3
Contributors are:
Who are from:
Tags
Data Science
D2L
Dive into Deep Learning @ D2L
Related
Mini-Batch Gradient Descent
Gradient Descent with Momentum
An overview of gradient descent optimization algorithms
Learning Rate Decay
Gradient Descent
AdaDelta (Deep Learning Optimization Algorithm)
Adam (Deep Learning Optimization Algorithm)
RMSprop (Deep Learning Optimization Algorithm)
AdaGrad (Deep Learning Optimization Algorithm)
Nesterov momentum (Deep Learning Optimization Algorithm)
Challenges with Deep Learning Optimizer Algorithms
Adam optimization algorithm
Difference between Adam and SGD
On a straight line, the function's derivative...
Gradient Descent
A crash course of derivatives
Learning Rate
Second Derivative
Hessian Matrix
Optimal Step Size according to Taylor Series Approximation
Lipschitz Continuous
Differentiation Rules
Derivatives of Common Functions
Chain Rule for Single-Variable Functions
Jacobian Matrix
Partial Derivative
Gradient of a Scalar-Valued Function with Respect to a Vector
Learn After
Gradient Descent Reference
Linear Regression and Gradient Descent
Numerical Approximation of Gradients
Gradient Checking
(Batch) Gradient Descent (Deep Learning Optimization Algorithm)
Gradient Descent Explained
Why Gradient descent might fail?
A Chat with Andrew on MLOps: From Model-centric to Data-centric AI
Big Data to Good Data: Andrew Ng Urges ML Community To Be More Data-Centric and Less Model-Centric
MLOps: Data-centric and Model-centric approaches
Critical Points
First-order Optimization Algorithm
Second-order Optimization Algorithm
Method of Steepest Descent
Second-Order Gradient Methods
Gradient Descent Explanation
Gradient Descent Variants
Notes about gradient descent
Suppose you have built a neural network. You decide to initialize the weights and biases to be zero. Which of the following statements is true?
Vanishing/exploding gradient
BERT Training Process
Objective Function
Distributed Training
The Problem with Constant Initialization