Question: Suppose you have a neural network with 3 layers. The first layer is the input layer with 20 neurons just encoding the 20 input binary

Suppose you have a neural network with 3 layers. The first layer is the input layer with 20 neurons just encoding the 20 input binary values (0 or 1). The second layer consists of 30 perceptrons each with their own weights and biases. The final layer is just one output perceptron with its weights and bias. The neurons in adjacent layers are fully connected. So this neural network computes a function y=f(x1,x2,,x20) where all variables take binary values. Assume that wx+b is never 0 for every neuron in the network and every possible input x. Is it possible to change the weights and biases of the perceptrons in this neural network in a systematic way so that the new neural network computes a new function z=h(x1,x2,,x20) such that h(x1,x2,,x20)=1- f(x1,x2,,x20) for all possible inputs x1,x2,,x20? If so, how?

Step by Step Solution

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Databases Questions!