Question: In a dropout layer, instead of zeroing out activations at test time, we multiply the weights by 1 p , where p is the probability

In a dropout layer, instead of zeroing out activations at test time, we multiply the weights by 1 p, where p is the probability that an activation is set to zero during training. Explain why the multiplication by 1 p is necessary for the neural network to make meaningful predictions.

Step by Step Solution

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Programming Questions!