Learn Before
Formula
Gradient of Objective Function with Respect to Hidden Layer Output
To continue backpropagation backwards towards the input layer, we calculate the gradient of the objective function with respect to the hidden layer output vector . By applying the chain rule through the output layer variable , we obtain: This operation successfully propagates the error gradient backward by multiplying it with the transpose of the output layer's weight matrix.
0
1
Updated 2026-05-06
Tags
D2L
Dive into Deep Learning @ D2L