Question: Consider a neural network for a binary classification which has one hidden layer as shown in the figure right. We use a linear activation function

Consider a neural network for a binary classification which has one hidden layer as shown in the figure right. We use a linear activation function h(z)= cz at hidden units and a ReLU activation function g(z)= max[0,z] at the output unit to learn the function for P(y =1| x, w) where x =(x1, x2) and w =(w1, w2,..., w7). What is the final classification boundary? (Note that your classifier predicts yhat =1 if the output g(z)>0.5, else yhat =0.)

Step by Step Solution

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Databases Questions!