Question: 3. Computational graph (no code involved) This question aims at checking your understanding on defining arbitrary network architectures and compute any derivative involved for optimization

3. Computational graph (no code involved) This question aims at checking your understanding on defining arbitrary network architectures and compute any derivative involved for optimization Consider a neural network with N input units, N output units, and K hidden units. The activations are computed as follows: W(1)x+ b(1) h=0(z) Z= y=x+ W(2)h + b(2) where o denotes the logistic function, applied elementwise. The cost involves a squared difference with the target s (with a 0.5 factor) and a regularization term that accounts for the dot product with respect to an external vector r. More concretely: E = R +S R=rh S = lly 5|12 a) Draw the computation graph relating x, z, h, y, R, S, and E. b) Derive the backpropagation equations for computing aE / W (1). To make things simpler, you may use o' to denote the derivative of the ReLU function. 3. Computational graph (no code involved) This question aims at checking your understanding on defining arbitrary network architectures and compute any derivative involved for optimization Consider a neural network with N input units, N output units, and K hidden units. The activations are computed as follows: W(1)x+ b(1) h=0(z) Z= y=x+ W(2)h + b(2) where o denotes the logistic function, applied elementwise. The cost involves a squared difference with the target s (with a 0.5 factor) and a regularization term that accounts for the dot product with respect to an external vector r. More concretely: E = R +S R=rh S = lly 5|12 a) Draw the computation graph relating x, z, h, y, R, S, and E. b) Derive the backpropagation equations for computing aE / W (1). To make things simpler, you may use o' to denote the derivative of the ReLU function
Step by Step Solution
There are 3 Steps involved in it
Get step-by-step solutions from verified subject matter experts
