Concept

Inverted Dropout Technique

In this version of dropout regularization, the activation vector is not only multiplied by a 0-1 dropout vector but also divided by the probability P (see parent node). This additional division, which is done at training time, preserves the expected value of the activations. This makes test time faster by eliminating scaling issues before the test phase.

0

2

Updated 2021-03-15

Tags

Data Science