Learn Before
Concept
Optimal Step Size according to Taylor Series Approximation
Denote the function as , is the gradient and is is the Hessian at . We calculate the new point . We can obtian that According to the above equation, the optimal step size when is positive is
0
1
Updated 2021-05-24
Tags
Data Science
Related
On a straight line, the function's derivative...
Gradient Descent
A crash course of derivatives
Learning Rate
Second Derivative
Hessian Matrix
Optimal Step Size according to Taylor Series Approximation
Lipschitz Continuous
Differentiation Rules
Derivatives of Common Functions
Chain Rule for Single-Variable Functions
Jacobian Matrix
Partial Derivative
Gradient of a Scalar-Valued Function with Respect to a Vector