Concept

Dropout in Deep Learning

  • Instead of training multiple models like bagging, dropout train the ensemble consisting of all subnetworks that can be formed by removing non-output units from an underlying base network. These models share parameters, with each model inheriting a different subset of parameters from parent networks.

  • We need to specify the probability of unit to be included. Typically, the input unit is included with probability 0.8 while that in hidden units is 0.5.

  • Prediction of ensemble is given by geometric mean.

  • Cheap computational cost and can be combined with other method or regularization.

0

1

Updated 2021-03-12

Contributors are:

Who are from:

Tags

Data Science