Learn Before
Concept
Inverted Dropout Technique
In this version of dropout regularization, the activation vector is not only multiplied by a 0-1 dropout vector but also divided by the probability P (see parent node). This additional division, which is done at training time, preserves the expected value of the activations. This makes test time faster by eliminating scaling issues before the test phase.
0
2
Updated 2021-03-15
Tags
Data Science