0% found this document useful (0 votes)
3 views7 pages

Ass4 Soln

The document contains an assignment for a Deep Learning course from IIT Kharagpur, consisting of 10 multiple-choice questions related to gradient descent, neural networks, and activation functions. Each question includes the correct answer and a detailed solution explaining the reasoning behind it. The topics covered include weight update rules, saddle points, logic functions, and the vanishing gradient problem.

Uploaded by

Revathi S
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
3 views7 pages

Ass4 Soln

The document contains an assignment for a Deep Learning course from IIT Kharagpur, consisting of 10 multiple-choice questions related to gradient descent, neural networks, and activation functions. Each question includes the correct answer and a detailed solution explaining the reasoning behind it. The topics covered include weight update rules, saddle points, logic functions, and the vanishing gradient problem.

Uploaded by

Revathi S
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 7

NPTEL Online Certification Courses

Indian Institute of Technology Kharagpur

Deep Learning
Assignment- Week 4
TYPE OF QUESTION: MCQ/MSQ
Number of questions: 10 Total mark: 10 X 1 = 10
______________________________________________________________________________

QUESTION 1:
A given cost function is of the form J( ) = 2 - ? What is the weight update rule for gradient
descent optimization at step t+1? Consider, =0.01 to be the learning rate.

a.
b.
c.
d.

Correct Answer: a

Detailed Solution:

So, weight update will be

______________________________________________________________________________

QUESTION 2:
Can you identify in which of the following graph gradient descent will not work correctly?

a. First figure
b. Second figure
c. First and second figure
NPTEL Online Certification Courses
Indian Institute of Technology Kharagpur

d. Fourth figure
Correct Answer: b

Detailed Solution:

This is a classic example of saddle point problem of gradient descent. In the second graph
gradient descent may get stuck in the saddle point.
______________________________________________________________________________

QUESTION 3:
From the following two figures can you identify which one corresponds to batch gradient
descent and which one to Stochastic gradient descent?

a. Graph-A: Batch gradient descent, Graph-B: Stochastic gradient descent


b. Graph-B: Batch gradient descent, Graph-A: Stochastic gradient descent
c. Graph-A: Batch gradient descent, Graph-B: Not Stochastic gradient descent
d. Graph-A: Not batch gradient descent, Graph-B: Not Stochastic gradient descent

Correct Answer: a

Detailed Solution:

The graph of cost vs epochs is quite smooth for batch gradient descent because we are
averaging over all the gradients of training data for a single step. The average cost over the
epochs in Stochastic gradient descent fluctuates because we are using one example at a
time.

______________________________________________________________________________
NPTEL Online Certification Courses
Indian Institute of Technology Kharagpur

QUESTION 4:
Suppose for a cost function as shown in graph below, in which point do you feel
magnitude of weight update will be more? is plotted along horizontal axis.

a. Red point (Point 1)


b. Green point (Point 2)
c. Yellow point (Point 3)
d. Red (Point 1) and yellow (Point 3) have same magnitude of weight update

Correct Answer: a

Detailed Solution:

Weight update is directly proportional to the magnitude of the gradient of the cost
function. In our case, . So, the weight update will be more for higher values of .

______________________________________________________________________________
NPTEL Online Certification Courses
Indian Institute of Technology Kharagpur

QUESTION 5:
Which logic function can be performed using a 2-layered Neural Network?

a. AND
b. OR
c. XOR
d. All

Correct Answer: d

Detailed Solution:

A two layer neural network can be used for any type logic Gate (linear or non linear)
implementation.
____________________________________________________________________________

QUESTION 6:
Let X and Y be two features to discriminate between two classes. The values and class labels of
the features are given hereunder. The minimum number of neuron-layers required to design
the neural network classifier

X Y #Class

0 2 Class-II

1 2 Class-I

2 2 Class-I

1 3 Class-I

1 -3 Class-II

a. 1
b. 2
c. 4
d. 5
Correct Answer: a.
NPTEL Online Certification Courses
Indian Institute of Technology Kharagpur

Detailed Solution:

Plot the feature points. They are linearly separable. Hence single layer is able to do the
classification task.

____________________________________________________________________________

QUESTION 7:
Which among the following options give the range for a logistic function?

a. -1 to 1
b. -1 to 0
c. 0 to 1
d. 0 to infinity

Correct Answer: c

Detailed Solution:

Refer to lectures, specifically the formula for logistic function.

______________________________________________________________________________

QUESTION 8:
The number of weights (including bias) to be learned by the neural network having 3 inputs and
2 classes and a hidden layer with 5 neurons is: (Assume we use 2 output nodes for 2 classes)

a. 12
b. 15
c. 25
d. 32
Correct Answer: d

Detailed Solution:

Please refer to lecture note week 4

(#input=3)+1(bias)x(#Hidden nodes=5) =(3+1)x5= 20 (#weights in 1st layer)


(#Hidden Nodes+1(bias))x(#classes=2)=(5+1)x2=12 (#weights in 2nd layer)

Hence, total weights= 20+12 =32


NPTEL Online Certification Courses
Indian Institute of Technology Kharagpur

______________________________________________________________________________

QUESTION 9:
For a XNOR function as given in the figure below, activation function of each node is given by:
. Consider and , what will be the output for the above
neural network?

a. 1.5
b. 2
c. 0
d. 1

Correct Answer: c

Detailed Solution:

Output of

Output of

Output of

So, the correct answer is c.

____________________________________________________________________________

QUESTION 10:
Which activation function is more prone to vanishing gradient problem?
NPTEL Online Certification Courses
Indian Institute of Technology Kharagpur

a. ReLU
b. Tanh

c. sigmoid

d. Threshold

Correct Answer: b

Detailed Solution:

Please refer to the lectures of week 4.

************END*******

You might also like