Key Concepts On Deep Neural Networks
Key Concepts On Deep Neural Networks
测验, 10 个问题
恭喜!您通过了! 下一项
1/1分
1。
What is the "cache" used for in our implementation of forward propagation and
backward propagation?
正确
Correct, the "cache" records values from the forward propagation units
and sends it to the backward propagation units because it is needed to
compute the chain rule derivatives.
1/1分
2。
Among the following, which ones are "hyperparameters"? (Check all that apply.)
[l]
bias vectors b
未选择的是正确的
正确
learning rate α
正确
[l]
weight matrices W
未选择的是正确的
正确
number of iterations
正确
1/1分
3。
Which of the following statements is true?
正确
1/1分
4。
Vectorization allows you to compute forward propagation in an L -layer neural
network without an explicit for-loop (or any other explicit iterative loop) over the
layers l=1, 2, …,L. True/False?
True
False
正确
Forward propagation propagates the input through the layers, although
for shallow networks we may just write all the lines (a[2] = g [2] (z [2] ) ,
[2]
z [2] = W [2] a[1] + b , ...) in a deeper network, we cannot avoid a for
loop iterating over the layers: (a[l] = g [l] (z [l] ) , z [l] = W [l] a[l−1] + b[l] ,
...).
1/1分
5。
Assume we store the values for n[l] in an array called layers, as follows:
layer_dims = [nx , 4,3,2,1]. So layer 1 has four hidden units, layer 2 has 3 hidden
units and so on. Which of the following for-loops will allow you to initialize the
parameters for the model?
1
2
3
for(i in range(1, len(layer_dims))):
parameter[‘W’ + str(i)] = np.random.randn(layers[i], layers[i-1])) * 0.01
Key concepts on Deep Neural Networks
parameter[‘b’ + str(i)] = np.random.randn(layers[i], 1) * 0.01 10/10 分 (100%)
测验, 10 个问题 正确
1/1分
6。
Consider the following neural network.
正确
Yes. As seen in lecture, the number of layers is counted as the number of
hidden layers + 1. The input and output layers are not counted as hidden
layers.
1/1分
7。
During forward propagation, in the forward function for a layer l you need to
know what is the activation function in a layer (Sigmoid, tanh, ReLU, etc.). During
backpropagation, the corresponding backward function also needs to know
what is the activation function for layer l , since the gradient depends on it.
True/False?
True
Key concepts正确on Deep Neural Networks 10/10 分 (100%)
测验, 10 个问题 Yes, as you've seen in the week 3 each activation has a different
derivative. Thus, during backpropagation you need to know which
activation was used in the forward propagation to be able to compute
the correct derivative.
False
1/1分
8。
There are certain functions with the following properties:
(i) To compute the function using a shallow network circuit, you will need a large
network (where we measure size by the number of logic gates in the network),
but (ii) To compute it using a deep network circuit, you need only an
exponentially smaller network. True/False?
True
正确
False
1/1分
9。
Consider the following 2 hidden layer neural network:
Key concepts on Deep Neural Networks 10/10 分 (100%)
测验, 10 个问题
Which of the following statements are True? (Check all that apply).
正确
[l]
Yes. More generally, the shape of W is (n[l] , n[l−1] ) .
正确
[l]
Yes. More generally, the shape of b is (n[l] , 1) .
未选择的是正确的
未选择的是正确的
正确
[l]
Yes. More generally, the shape of W is (n[l] , n[l−1] ) .
will have shape (1, 1)
Key concepts on Deep Neural Networks 10/10 分 (100%)
测验, 10 个问题
b[2]
未选择的是正确的
未选择的是正确的
正确
[l]
Yes. More generally, the shape of b is (n[l] , 1) .
未选择的是正确的
正确
[l]
Yes. More generally, the shape of b is (n[l] , 1) .
正确
[l]
Yes. More generally, the shape of W is (n[l] , n[l−1] ) .
未选择的是正确的
1/1分
10。
Whereas the previous question used a specific network, in the general case what
is the dimension of W^{[l]}, the weight matrix associated with layer l ?