Question: In a dropout layer, instead of zeroing out activations at test time, we multiply the weights by 1 p , where p is the probability
In a dropout layer, instead of zeroing out activations at test time, we multiply the weights by p where p is the probability that an activation is set to zero during training. Explain why the multiplication by p is necessary for the neural network to make meaningful predictions.
Step by Step Solution
There are 3 Steps involved in it
1 Expert Approved Answer
Step: 1 Unlock
Question Has Been Solved by an Expert!
Get step-by-step solutions from verified subject matter experts
Step: 2 Unlock
Step: 3 Unlock
