Quiz 3 in lecture 3

Which of the following neural networks are examples of a feed-forward neural network?

第 2 个问题

2。第 2 个问题

Consider a neural network with only one training case with input \mathbf{x} = (x_1, x_2, \ldots, x_n)^\topx=(x1,x2,,xn)and correct output tt. There is only one output neuron, which is linear, i.e. y = \mathbf{w}^\top\mathbf{x}y=wx(notice that there are no biases). The loss function is squared error. The network has no hidden units, so the inputs are directly connected to the output neuron with weights \mathbf{w} = (w_1, w_2, \ldots, w_n)^\topw=(w1,w2,,wn). We're in the process of training the neural network with the backpropagation algorithm. What will the algorithm add to w_iwi for the next iteration if we use a step size (also known as a learning rate) of \epsilonϵ?

第 3 个问题

3。第 3 个问题

Suppose we have a set of examples and Brian comes in and duplicates every example, then randomly reorders the examples. We now have twice as many examples, but no more information about the problem than we had before. If we do not remove the duplicate entries, which one of the following methods will not be affected by this change, in terms of the computer time (time in seconds, for example) it takes to come close to convergence?

第 4 个问题

4。第 4 个问题

Consider a linear output unit versus a logistic output unit for a feed-forward network with no hidden layer shown below. The network has a set of inputs xx and an output neuron yyconnected to the input by weights ww and bias bb.

We're using the squared error cost function even though the task that we care about, in the end, is binary classification. At training time, the target output values are 11 (for one class) and 00 (for the other class). At test time we will use the classifier to make decisions in the standard way: the class of an input xx according to our model after training is as follows:

\text{class of }x=

{1 if wTx+b00 otherwise
class of x={1 if wTx+b00 otherwise

Note that we will be training the network using yy, but that the decision rule shown above will be the same at test time, regardless of the type of output neuron we use for training.

Which of the following statements is true?

第 5 个问题

5。第 5 个问题

Consider a neural network with one layer of logistic hidden units (intended to be fully connected to the input units) and a linear output unit. Suppose there are nn input units and mm hidden units. Which of the following statements are true? Check all that apply.

第 6 个问题

6。第 6 个问题

Brian wants to make his feed-forward network (with no hidden units) using a linearoutput neuron more powerful. He decides to combine the predictions of two networks by averaging them. The first network has weights w_1w1 and the second network has weights w_2w2. The predictions of this network for an example xx are therefore:

y=12wT1x+12wT2x

Can we get the exact same predictions as this combination of networks by using a single feed-forward network (again with no hidden units) using a linear output neuron and weights w_3=\frac{1}{2}(w_1+w_2)w3=21(w1+w2)?




第 6 个问题

Brian wants to make his feed-forward network (with no hidden units) using a logisticoutput neuron more powerful. He decides to combine the predictions of two networks by averaging them. The first network has weights w_1w1 and the second network has weights w_2w2. The predictions of this network for an example xx are therefore:

y=1211+ez1+1211+ez2 with z_1=w_1^Txz1=w1Tx and z_2=w_2^Txz2=w2Tx.

Can we get the exact same predictions as this combination of networks by using a single feed-forward network (again with no hidden units) using a logistic output neuron and weights w_3=\frac{1}{2}(w_1+w_2)w3=21(w1+w2)?


第 6 个问题

Brian wants to make his feed-forward network (with no hidden units) using a logisticoutput neuron more powerful. He decides to combine the predictions of two networks by averaging them. The first network has weights w_1w1 and the second network has weights w_2w2. The predictions of this network for an example xx are therefore:

y=1211+ez1+1211+ez2 with z_1=w_1^Txz1=w1Tx and z_2=w_2^Txz2=w2Tx.

Can we get the exact same predictions as this combination of networks by using a single feed-forward network (again with no hidden units) using a logistic output neuron and weights w_3=\frac{1}{2}(w_1+w_2)w3=21(w1+w2)?

猜你喜欢

转载自blog.csdn.net/sophiecxt/article/details/80424377