Deep Learning MCQA
Deep Learning MCQA
MCQA
1. Supervised learning and unsupervised clustering both require at least one
a. hidden attribute.
b. output attribute.
c. input attribute.
d. categorical attribute.
7. How many individuals in the class life insurance= no have credit card
insurance and are less than 30 years old?
a. 63
b. 70
c. 30
d. 27
8. Which statement is true about neural network and linear regression models?
a. Both models require input attributes to be numeric.
b. Both models require numeric attributes to range between 0 and 1.
c. The output of both models is a categorical attribute value.
d. Both techniques build models whose output is determined by a linear sum of
weighted input attribute values.
e. More than one of a,b,c or d is true.
11 Unlike traditional production rules, association rules
a. allow the same variable to be an input attribute in one rule and an output
attribute in another rule.
b. allow more than one input attribute in a single rule.
c. require input attributes to take on numeric values.
d. require each rule to have exactly one categorical output attribute.
13. The average positive difference between computed and desired outcome values.
a. root mean squared error
b. mean squared error
c. mean absolute error
d. mean positive error
20 With a Kohonen network, the output layer node that wins an input instance is
rewarded by having
a. a higher probability of winning the next training instance to be presented.
b. its connect weights modified to more closely match those of the input
instance.
c. its connection weights modified to more closey match those of its neighbors.
d. neighoring connection weights modified to become less similar to its own
connection weights.
24 During backpropagation training, the purpose of the delta rule is to make weight
adjustments so as to
a. minimize the number of times the training data must pass through the network.
b. minimize the number of times the test data must pass through the network.
c. minimize the sum of absolute differences between computed and actual outputs.
d. minimize the sum of squared error differences between computed and actual
output.
26 Two classes each of which is represented by the same pair of numeric attributes
are linearly separable if
a. at least one of the pairs of attributes shows a curvilinear relationship between the
classes.
b. at least one of the pairs of attributes shows a high positive correlation between the
classes.
c. at least one of the pairs of attributes shows a high positive correlation between the
classes.
d. a straight line partitions the instances of the two classes.
27 The test set accuracy of a backpropagation neural network can often be improved
by
a. increasing the number of epochs used to train the network.
b. decreasing the number of hidden layer nodes.
c. increasing the learning rate.
d. decreasing the number of hidden layers.
28 This type of supervised network architecture does not contain a hidden layer.
a. backpropagation
b. perceptron
c. self-organizing map
d. genetic
29 The total delta measures the total absolute change in network connection weights
for each pass of the training data through a neural network. This value is most
often used to determine the convergence of a
a. perceptron network.
b. feed-forward network.
c. backpropagation network.
d. self-organizing network.
34 This technique associates a conditional probability value with each data instance.
a. linear regression
b. logistic regression
c. simple regression
d. multiple linear regression
38 Choose the options that are incorrect regarding machine learning (ML) and
artificial intelligence (AI),
(A) ML is an alternate way of programming intelligent machines.
(B) ML and AI have very different goals.
(C) ML is a set of techniques that turns a dataset into a software.
(D) AI is a software that can emulate the human mind.
42 K-fold cross-validation is
(A) linear in K
(B) quadratic in K
(C) cubic in K
(D) exponential in K
43 You observe the following while fitting a linear regression to the data: As
you increase the amount of training data, the test error decreases and the
training error increases. The train error is quite low (almost what you expect
it to), while the test error is much higher than the train error.
What do you think is the main reason behind this behavior. Choose the
most probable option.
(A) High variance
(B) High model bias
(C) High estimation bias
(D) None of the above
44 Adding more basis functions in a linear model... (pick the most probably
Option)
(A) Decreases model bias
(B) Decreases estimation bias
(C) Decreases variance
(D) Doesn't affect bias and variance
45.The number of nodes in the input layer is 10 and the hidden layer is 5. The maximum number
of connections from the input layer to the hidden layer are
A) 50
B) Less than 50
C) More than 50
D) It is an arbitrary value
46.Which of the following functions can be used as an activation function in the output layer
if we wish to predict the probabilities of n classes (p1, p2..pk) such that sum of p over all n
equals to 1?
A) Softmax
B) ReLu
C) Sigmoid
D) Tanh
47.Assume a simple MLP model with 3 neurons and inputs= 1,2,3. The weights to the input
neurons are 4,5 and 6 respectively. Assume the activation function is a linear constant value of 3.
A) 32
B) 643
C) 96
D) 48
48. In the neural network, every parameter can have their different learning rate.
A) TRUE
B) FALSE
49.In the neural network, every parameter can have their different learning rate.
A) TRUE
B) FALSE
50The red curve above denotes training accuracy with respect to each epoch in a deep
learning algorithm. Both the green and blue curves denote validation accuracy.
A) Green Curve
B) Blue Curve
A) Both 1 and 2
B) Both 1 and 3
C) Both 2 and 3
D) All 1, 2 and 3
54.IN Neural network The input from Input layer is then feed into the______.
A. Input layer
B. Output layer
C. Hidden layer
D. None of these
55.____________ computes the output volume by computing dot product between all filters and
image patch.
A. Input Layer
B. Convolution Layer
C. Activation Function Layer
D. Pool Layer
56.________is a pooling operation that selects the maximum element from the region of the
feature map covered by the filter.
A. Max Pooling
B. Average Pooling
C. Global pooling
D. None of these
57.Recurrent Neural Networks are best suited for Text Processing.
True
False
58.What does LSTM stand for?
Long Short Term Memory
Least Squares Term Memory
Least Square Time Mean
Long Short Threshold Memory
Answer:-Long Short Term Memory
59.What is the difference between the actual output and generated output known as?
Output Modulus
Accuracy
Cost
Output Difference
Answer:-Cost
60. Recurrent Neural Networks are best suited for Text Processing.
True
False
Answer:-True
66.Name the component of a Neural Network where the true value of the input is not observed.
Hidden Layer
Gradient Descent
Activation Function
Output Layer
Answer:- Hidden Layer
68.Neural Networks Algorithms are inspired from the structure and functioning of the Human
Biological Neuron.
False
True
Answer:- True
69. In a Neural Network, all the edges and nodes have the same Weight and Bias values.
True
False
Answer:- False
73. A Shallow Neural Network has only one hidden layer between Input and Output layers.
False
True
Answer:- True
74. Support Vector Machines, Naive Bayes and Logistic Regression are used for solving
___________________ problems.
Clustering
Classification
Regression
Time Series
Answer:- Classification
75. The rate at which cost changes with respect to weight or bias is called __________________.
Derivative
Gradient
Rate of Change
Loss
77. All the Visible Layers in a Restricted Boltzmannn Machine are connected to each other.
True
False
Answer:- False
78.All the neurons in a convolution layer have different Weights and Biases.
True
False
Answer:- False
79.What is the method to overcome the Decay of Information through time in RNN known as?
Back Propagation
Gradient Descent
Activation
Gating
Answer:- Gating
80. Recurrent Network can input Sequence of Data Points and Produce a Sequence of Output.
False
True
Answer:- True
82. Restricted Boltzmann Machine expects the data to be labeled for Training.
False
True
Answer:- False
89.The rate at which cost changes with respect to weight or bias is called __________________.
Derivative
Gradient
Rate of Change
Loss
Answer:- Gradient