Question: Suppose we are running gradient descent on the neural network above. We are trying to minimize the upstream loss L L using the learning rate
Suppose we are running gradient descent on the neural network above. We are trying to minimize the upstream loss L L using the learning rate . Given the upstream gradient L y y L and two partial derivatives that you calculated in the previous part ( y z z y and z w 0 w 0 z ), determine the gradient update rule for w 0 w 0 . Write your final answer in the form of " w 0 f w 0 f, where f f represents the expression for the update rule
Step by Step Solution
There are 3 Steps involved in it
1 Expert Approved Answer
Step: 1 Unlock
Question Has Been Solved by an Expert!
Get step-by-step solutions from verified subject matter experts
Step: 2 Unlock
Step: 3 Unlock
