0% found this document useful (0 votes)
11 views

Neural Network and Deep Learning

NNDL Paper for ASTU

Uploaded by

kaustavkakoty4
Copyright
© © All Rights Reserved
Available Formats
Download as PDF or read online on Scribd
0% found this document useful (0 votes)
11 views

Neural Network and Deep Learning

NNDL Paper for ASTU

Uploaded by

kaustavkakoty4
Copyright
© © All Rights Reserved
Available Formats
Download as PDF or read online on Scribd
You are on page 1/ 4
Total No. CSE 1818 PE 63 jofefr- Roll.No, ofeandidata, | | i | | of printed pages = 3 o 2022 B.Tech. (CSE) 8** Semester End-Term Examination CSE NEURAL NETWORK AND DEEP LEARNING (New Regulation 2017-18 & New Syllabus 2018-19) Full Marks -70 Time — Three hours ‘The figures in the margin indicate full marks for the questions. 1. Answer the following questions : (0x1 =10) (In CNN, which layer is responsible to perform dimensionality reduction? (a) Convolutional Layer (b) Fully Connected Layer (©) Pooling layer (@ Allof the above (ii) Does the size of the feature map reduce upon applying a filter of size 1 x 1? ii) 0) @) (a) Yes 2 (b) No The larger the Strides the ——————— the feature map. (a) Larger (b) Smailer (©) Size remains same @) None of the above Which of the following is/are limitations of deep learning? (a) Data labeling (>) Obtain huge training datasets (© Both (a) and () @ None of the above In which of the following applications can we use deep learning to solve the problem? (a) Protein structure prediction (&) Prediction of chemical reactions (© Detection of exotic particles (@) Allof the above [Turn over (vi) In an MLP, the number of nodes in the input layer is 10 and the hidden layer is 5. The maximum number of connections from the input layer to the hidden layer are (a) 50 () less than 50 (© more than 50 (@) It is an arbitrary value (vii) Which of the following neural network training challenge can be solved using batch normalization? (a) Over fitting (©) Restrict activations to become too high or low (©) Training is too slow (d@) Both (b) and (c) OP pater (viii) In a simple MLP model with 8 neurons in the inpiit layer, 5 neurons in the hidden layer and 1 neuron in the output layer. What isthe size of the weight matrices between input-hidden layer and hidden-output layer? (a) [1x5], [5x8] ) (8x5), [5x1] © [6 8), («1 @) [8x 5], [1x 5) (ix) Which of the following functions can be used as an activation function in the output layer if we wish to predict the probabilities of n classes (p1, p2..pk) such that sum of p over all n equals to I? (@) Softmax (b) ReLu © Sigmoid @) Tanh (x) The C; curve above denotes tracing accuracy w.r.t. each epoch in a deep learning algorithm. Both the C2 and Cs curves denote validation accuracy. Which of these indicate overfiting? faccuracy ay te G ‘epoch (a) Czcurve () Cs curve CSE 1818 PE 63 2 Answer the following questions : (Attempt only four questions) : 2. (a) Write a simple Multilayer Perceptron program or a pseudocode(without using the MLP library function) having two input nodes, one hidden layer (with 4 nodes in it) and two nodes at the output layer. (12+3= 15) (b) What is over-fitting and under-fitting? 3. (a) Explain the gradient descent algorithm. (8+ 7 = 15) (b) What is the need of an activation function? Distinguish between sigmoid, rectified linear unit and softmax activation function. 4. (a) Explain in detail, how to perform Principal Component Analysis (PCA). (8+7=15) (b) What are autoencoders used for? How to regularize autoencoders? 5. (a) Below is a diagram of a small convolutional neural network that converts a 13 x 13 image into 4 output values. The network has the following layers/operations from input to output: convolution with 3 filters, max pooling, ReLu, and finally a fully-connected layer. For this network we will not be using any bias/offset parameters. Please answer the following questions about'this network. (8+ 7=16) 13x13 3@10x10 # Convolution. Max Pooling 3 Filters 4x4 2x2 Stride 1 Stride 2 @ How many weights in the convolutional layer do we need to learn? (i) How many Rel.u operations are performed on the forward pass? (ii) How many weights do we need to learn for the entire network? (iv) Ifthe input image has been converted into a matrix of size 28 x 28 and a kernel/filter of size 7 x 7 with a stride of 1 is used, what will be the size of the convoluted matrix? (») What is the significance of a Pooling layer? What are the different hyperparameters of a pooling layer? Explain the different types of pooling layer. CSE 1818 PE 63 be 3 6. (a) (b) © @) What is the meaning of valid padding and same padding in CNN? (4+4+4+3=16) What is Dropout and Batch Normalization? What are Vanishing and Exploding Gradients? What are Hyperparameters? 7. Write short notes on (any three) : (@x5=15) (a) ) © @ ©) © () Bias-Variance Tradeoff Batch Normalization Ensemble methods Long Short-Term Memory * Encoder Decoder Models Covariance and Correlation Pre-trained network CSE 1818 PE 63 4

You might also like