Learn Before
Concept

Mini-Batch Gradient Descent Algorithm

for t = 1, 2,...N: (N is the number of mini-batches)

  • Forward propagate on X{t}X^{\{ t\}}
  • Compute cost function J{t}J^{\{ t\}}
  • Backpropagate to compute gradients wrt J{t}J^{\{ t\}} (using X{t}X^{\{ t\}},Y{t}Y^{\{ t\}})
  • W[l]=W[l]αdW[l],b[l]=b[l]αdb[l]W^{[l]} =W^{[l]}-\alpha dW^{[l]}, b^{[l]} = b^{[l]}-\alpha db^{[l]}

This is one pass through your training set using mini-batch gradient descent. It is also called doing one epoch of training.

0

2

Updated 2020-11-16

Tags

Data Science