Doubt about the implementation of inverted dropout

Hi, @jaylodha.

By dropping some units during training you are changing the expected values of the activations with respect to test time. To compensate for this, you can either scale down the weights at test time (multiplying by keep_probs) , or scale up the activations at training time (dividing by keep_probs).

You have all the details here.

Happy learning! :slight_smile:

4 Likes