0% found this document useful (0 votes)
115 views

Key Concepts On Deep Neural Networks

1. The cache is used to pass variables computed during forward propagation to the corresponding backward propagation step. It contains useful values for backward propagation to compute derivatives. 2. Hyperparameters include the size of the hidden layers, learning rate, number of layers in the neural network, and number of iterations. 3. Vectorization allows you to compute forward propagation in an L-layer neural network without an explicit for-loop over the layers.

Uploaded by

Shivam yadav
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
115 views

Key Concepts On Deep Neural Networks

1. The cache is used to pass variables computed during forward propagation to the corresponding backward propagation step. It contains useful values for backward propagation to compute derivatives. 2. Hyperparameters include the size of the hidden layers, learning rate, number of layers in the neural network, and number of iterations. 3. Vectorization allows you to compute forward propagation in an L-layer neural network without an explicit for-loop over the layers.

Uploaded by

Shivam yadav
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 8

Key concepts on Deep Neural Networks 10/10 分 (100%)

测验, 10 个问题

 恭喜!您通过了! 下一项

 1/1分

1。
What is the "cache" used for in our implementation of forward propagation and
backward propagation?

We use it to pass variables computed during forward propagation to


the corresponding backward propagation step. It contains useful
values for backward propagation to compute derivatives.

正确
Correct, the "cache" records values from the forward propagation units
and sends it to the backward propagation units because it is needed to
compute the chain rule derivatives.

It is used to keep track of the hyperparameters that we are searching


over, to speed up computation.

It is used to cache the intermediate values of the cost function during


training.

We use it to pass variables computed during backward propagation to


the corresponding forward propagation step. It contains useful values
for forward propagation to compute activations.

 1/1分

2。
Among the following, which ones are "hyperparameters"? (Check all that apply.)

activation values a[l]


Key concepts未选择的是正确的
on Deep Neural Networks 10/10 分 (100%)
测验, 10 个问题

[l]
bias vectors b

未选择的是正确的

 size of the hidden layers n[l]

正确

 learning rate α

正确

[l]
weight matrices W

未选择的是正确的

 number of layers L in the neural network

正确

 number of iterations

正确

 1/1分

3。
Which of the following statements is true?

The deeper layers of a neural network are typically computing more


complex features of the input than the earlier layers.

正确

The earlier layers of a neural network are typically computing more


complex features of the input than the deeper layers.
Key concepts on Deep Neural Networks 10/10 分 (100%)
测验, 10 个问题

 1/1分

4。
Vectorization allows you to compute forward propagation in an L -layer neural
network without an explicit for-loop (or any other explicit iterative loop) over the
layers l=1, 2, …,L. True/False?

True

False

正确
Forward propagation propagates the input through the layers, although
for shallow networks we may just write all the lines (a[2] = g [2] (z [2] ) ,
[2]
z [2] = W [2] a[1] + b , ...) in a deeper network, we cannot avoid a for
loop iterating over the layers: (a[l] = g [l] (z [l] ) , z [l] = W [l] a[l−1] + b[l] ,
...).

 1/1分

5。
Assume we store the values for n[l] in an array called layers, as follows:
layer_dims = [nx , 4,3,2,1]. So layer 1 has four hidden units, layer 2 has 3 hidden
units and so on. Which of the following for-loops will allow you to initialize the
parameters for the model?

1 for(i in range(1, len(layer_dims)/2)):


2 parameter[‘W’ + str(i)] = np.random.randn(layers[i], layers[i-1])) * 0.01
3 parameter[‘b’ + str(i)] = np.random.randn(layers[i], 1) * 0.01

1 for(i in range(1, len(layer_dims)/2)):


2 parameter[‘W’ + str(i)] = np.random.randn(layers[i], layers[i-1])) * 0.01
3 parameter[‘b’ + str(i)] = np.random.randn(layers[i-1], 1) * 0.01

1 for(i in range(1, len(layer_dims))):


2 parameter[‘W’ + str(i)] = np.random.randn(layers[i-1], layers[i])) * 0.01
3 parameter[‘b’ + str(i)] = np.random.randn(layers[i], 1) * 0.01

1
2
3
for(i in range(1, len(layer_dims))):
parameter[‘W’ + str(i)] = np.random.randn(layers[i], layers[i-1])) * 0.01
Key concepts on Deep Neural Networks
parameter[‘b’ + str(i)] = np.random.randn(layers[i], 1) * 0.01 10/10 分 (100%)
测验, 10 个问题 正确

 1/1分

6。
Consider the following neural network.

How many layers does this network have?

The number of layers L is 4. The number of hidden layers is 3.

正确
Yes. As seen in lecture, the number of layers is counted as the number of
hidden layers + 1. The input and output layers are not counted as hidden
layers.

The number of layers L is 3. The number of hidden layers is 3.

The number of layers L is 4. The number of hidden layers is 4.

The number of layers L is 5. The number of hidden layers is 4.

 1/1分

7。
During forward propagation, in the forward function for a layer l you need to
know what is the activation function in a layer (Sigmoid, tanh, ReLU, etc.). During
backpropagation, the corresponding backward function also needs to know
what is the activation function for layer l , since the gradient depends on it.
True/False?

True
Key concepts正确on Deep Neural Networks 10/10 分 (100%)
测验, 10 个问题 Yes, as you've seen in the week 3 each activation has a different
derivative. Thus, during backpropagation you need to know which
activation was used in the forward propagation to be able to compute
the correct derivative.

False

 1/1分

8。
There are certain functions with the following properties:

(i) To compute the function using a shallow network circuit, you will need a large
network (where we measure size by the number of logic gates in the network),
but (ii) To compute it using a deep network circuit, you need only an
exponentially smaller network. True/False?

True

正确

False

 1/1分

9。
Consider the following 2 hidden layer neural network:
Key concepts on Deep Neural Networks 10/10 分 (100%)
测验, 10 个问题

Which of the following statements are True? (Check all that apply).

 W [1] will have shape (4, 4)

正确
[l]
Yes. More generally, the shape of W is (n[l] , n[l−1] ) .

 b[1] will have shape (4, 1)

正确
[l]
Yes. More generally, the shape of b is (n[l] , 1) .

W [1] will have shape (3, 4)

未选择的是正确的

b[1] will have shape (3, 1)

未选择的是正确的

 W [2] will have shape (3, 4)

正确
[l]
Yes. More generally, the shape of W is (n[l] , n[l−1] ) .
will have shape (1, 1)
Key concepts on Deep Neural Networks 10/10 分 (100%)
测验, 10 个问题
b[2]
未选择的是正确的

W [2] will have shape (3, 1)

未选择的是正确的

 b[2] will have shape (3, 1)

正确
[l]
Yes. More generally, the shape of b is (n[l] , 1) .

W [3] will have shape (3, 1)

未选择的是正确的

 b[3] will have shape (1, 1)

正确
[l]
Yes. More generally, the shape of b is (n[l] , 1) .

 W [3] will have shape (1, 3)

正确
[l]
Yes. More generally, the shape of W is (n[l] , n[l−1] ) .

b[3] will have shape (3, 1)

未选择的是正确的

 1/1分

10。
Whereas the previous question used a specific network, in the general case what
is the dimension of W^{[l]}, the weight matrix associated with layer l ?

W [l] has shape (n[l] , n[l−1] )


Key concepts正确on Deep Neural Networks 10/10 分 (100%)
测验, 10 个问题 True

W [l] has shape (n[l] , n[l+1] )

W [l] has shape (n[l−1] , n[l] )

W [l] has shape (n[l+1] , n[l] )

  

You might also like