MCQ Dlei
MCQ Dlei
Question 1
Correct
The complexity in ANN depends upon the type of functional mapping, accurate and sufficient training data acquired, their effective way of
representation. Weighted sum in ANN are referred to as
a. Activation
b. Link
c. output
d. Input
Question 2
Correct
Artificial Neural Network (ANN) are type of Machine Learning model that are inspired by the structure and function of the human brain. In
________________ type of ANN information passed till it reaches the output.
Incorrect
ANN used for Classification, Clustering and Pattern recognition. A single layer feedforward neural network with preprocessing is perceptron.
If the change in weight vector is represented by Δ w ij what does it mean?
a. Describes the change in weight vector for ith processing unit, taking input vector jth into account
b. Describes the change in weight vector for jth processing unit, taking input vector ith into account
c. Describes the change in weight vector for jth and ith processing unit
The correct answer is: Describes the change in weight vector for ith processing unit, taking input vector jth into account
Question 4
Correct
Perceptron learning rule states that the algorithm would automatically learn the optimal weight coefficient. To find minimum or maximum
of a function, we set gradient to zero because which of the follwing
b. None of all
The correct answer is: the value of gradient at extrema of a function is always zero
Question 5
Correct
a. NONE
b. ALL
c. A AND B
d. ONLY B
Correct
The back propagation law is also known as generalized delta rule. Error in output is propagated backwards only to determine weight
updates. There is no feedback of signal at any stage. The objective of back propagation algorithm is _______________
a. To develop learning algorithm for multilayer feedforward neural network, so that network can be trained to capture the
mapping implicitly
The correct answer is: To develop learning algorithm for multilayer feedforward neural network, so that network can be trained to capture
the mapping implicitly
Question 7
Correct
A Gradient Descent algorithm is the key to minimizing the loss function and achieving our target, which is to predict close to the original
value. Does the back propagation learning is based on Gradient Descent algorithm along error surface
a. YES
b. NO
c. CAN'T BE SAID
Question 8
Correct
Outliers are those data points that are significantly different from the rest of the dataset. Which statement about outliers is true?
c. Outliers should be part of of the test dataset but should not be present in the training dataset.
d. Outliers should be part of the training dataset but should not be present in the test data set
The correct answer is: The nature of the problem determines how outliers are used.
Question 9
Correct
Supervised learning is the type of machine learning that uses labeled data to train a model to make predictions. unsupervised learning on
the other hand, doesn't require labeled data and instead uses algorithms to find patterns and relationships in the data. What is the
purpose of regularization in deep learning?
Question 10
Correct
A validation data set and test data set are not the same thing. A validation data set is usually necessary in machine/ deep learning to
prevent overfitting. A validation data set is used to___________ while a test set is used to ____________________.
a. A validation set is used to evaluate the performance of a model during training, while a test set is used to evaluate its
performance after training.
d. A validation set is used to tune the hyperparameters of a mode, while a test set is used to evaluate its performance
The correct answer is: A validation set is used to evaluate the performance of a model during training, while a test set is used to evaluate its
performance after training.
Question 11
Incorrect
The Akaike information criterion is a measure of model complexity used in machine /deep learning and statistics to compare the
performance of different models. It takes into account both the goodness of fit and the number of parameters in the model, and penalizes
models with more parameters. What is the purpose of data augmentation in deep learning?
Correct
A hyperparameter is a parameter that is set before the training process begins and cannot be learned directly from the data. Examples of
hyperparameters include________ which determines the step size taken during gradient descent optimization and the number of hidden
units in a neural network which controls its capacity and complexity.
c. Accuracy
d. Learning rate
Question 13
Correct
Non-linear relationships are common problems in regression tasks, where response variable does not vary linearly with the predictors. There
are several approaches to resolve these problems. Which of the following is the common approach to solve regression problem with
nonlinear relationships?
a. Polynomial regression
b. Linear Regression
c. Neural Network
d. all of them
Question 14
Correct
In PCA the number of input dimension are equal to principal components. PCA can be used for projecting and visualizing data in lower
dimensions. In following ___________________type of feature selection method we start with empty feature set
a. None of them
b. In both
Incorrect
The PCA method is used for feature extraction. The output of training process in machine learning is a machine learning model.
Dimensionality reduction algorithms are one of the possible ways to reduce the computation time required to build a model. Which of the
following techniques would perform better for dimensions of a data set?
b. None of these
The correct answer is: Removing columns which have too many missing values.
Question 16
Correct
Penalized logistic regression imposes a penalty to the logistic model for having too many variables. This results in shrinking the coefficient
of the less contributive variables towards___________
a. zero
b. one
c. Infinity
Question 17
Correct
Ensemble methods are techniques that create multiple models and then combine them to produce improved results. Ensemble learning
techniques employ ______________________________________
a. Reduce accuracy of classification models or to increase mean absolute error for regression models.
b. To increase the accuracy of classification models or to reduce mean absolute error for regression models.
c. To increase the accuracy of classification models or to increase mean absolute error for regression models.
d. To reduce the accuracy of classification models or to reduce absolute error for regression models.
The correct answer is: To increase the accuracy of classification models or to reduce mean absolute error for regression models.
Question 18
Correct
Current Neural Network based feature selection methods employs a simple autoencoder to perform feature selection based on
reconstruction error. For feature selection it is necessary to have a simple network so that error can be easily back propagated. Typically
Feature selection method consist of
1. Subset Generation
2. Subset evaluation
3. stopping criteria
4. Result validation
c. All four
d. No steps required
Question 19
Correct
In deep neural network, what is the role of nonlinear activation functions such as sigmoid, Tanh and ReLU?
The correct answer is: They help to learn nonlinear decision boundaries.
Question 20
Correct
Classification is a type of supervised learning where a target feature, which is of categorial type, is predicted for test data on the basis of
information imparted by the training data. The target categorial feature is known as
a. variable
b. object
c. class
d. method
◄ lab assignment 3
Jump to...
Question 1
Not answered
A. One way to implement undercomplete autoencoder is to constrain the number of nodes present in hidden layers of the neural network.
B. Autoencoders are capable of learning a continuous nonintersecting surface.
Question 2
Correct
Suppose you have built a neural network. You decide to initialize the weights and biases to be zero. Which of the following statements are
true?
1. Each neuron in the first hidden layer will perform the same computation. So even after multiple iterations of gradient descent each
neuron in the layer will be computing the same thing as other neurons.
2. Each neuron in the hidden layer will perform the same computation in the first iteration. But after one iteration of gradient descent they
will learn to compute different things because we have broken symmetry.
3. Each neuron in the first hidden layer will compute the same thing, but neurons in different layers will compute different things.
4. The first hidden layers neurons will perform different computations from each other even in the first iteration; their parameters will thus
keep evolving in their own way.
a. Statement 3 is true.
b. Statement 2 is true.
c. Statement 4 is true.
d. Statement 1 is true.
Correct
b. Use pervious inputs to find the next output according to the training set.
c. Use a loop between inputs and outputs in order to achieve the better prediction.
d. Use loops between the most important features to predict next output.
The correct answer is: Use pervious inputs to find the next output according to the training set.
Question 4
Correct
Padding on the input volume with zeros in such a way that the convolution layer does not alter the spatial dimensions of the input
The correct answer is: Keep the spatial dimensions of input intact.
Question 5
Incorrect
Correct
You are training an RNN and find that you are weights and activation are all taking on the value of " not a number". Which of these is the
most likely cause of this problem?
a. ReLU activation function used to compute with too large input value.
b. Sigmoid activation function used to compute with too large input value.
c. Exploding gradient
d. Vanishing gradient
Question 7
Incorrect
Benefits of Pooling
Question 8
Incorrect
The correct answer is: LSTM solves the exploding gradients issues in RNN.
Question 9
Correct
For a Input image 6*6, stride size 1 and convolution filter size 3*3....... output Feature map size is
b. 3*3
c. 4*4
d. 6*6
Question 10
Incorrect
GATED RECURRENT UNIT (GRU) network is one of the most famous alternative architectures.
1. It combines the forget and input gates into a single update gate.
2. it also merges the cell state and hidden state and makes some other changes.
3. Generates an output at each time step and this output used to train the network
Question 11
Correct
a. Leaky ReLU
c. Tanh
d. ReLU
Incorrect
Convolution leverages three important ideas that can help improve a Machine/ Deep learning system
The parameter sharing used by the convolution operation menace that rather than learning a separate set of parameters for every location
we learn only one set.
The correct answer is: 1st and 3rd points are true.
Question 13
Correct
We can reduce the exploding gradients problem by clipping the gradient during back propagation so that they never exceed some
threshold. This is called___________
a. Gradient Vanishing
b. Gradient Clipping
c. Gradient Overacting
d. Gradient Cutting
Correct
The Gated Recurrent Unit (GRU) cell was proposed by Kyunghyun Cho et.al. in the year _________
a. 2002
b. 2016
c. 2014
d. 2019
Question 15
Incorrect
1. The LSTM modify the state only through linear interactions: information flows smoothly across time.
2. The LSTM modify the state only through nonlinear interactions: information flows randomly across time.
3. LSTM protect and control the information in the cell through gates.
4. LSTM gates are implemented by tanh and a pointwise multiplication.
Correct
The denoising encourages the encoder to keep important information but forget about spurious information about the input.
A. The loss is between the original input and the reconstruction from a noisy version of the input.
Question 17
Incorrect
b. hidden function
c. visible function
Question 18
Correct
You are building a binary classifier for recognizing cucumbers (y=1) Vs watermelons (y=0). Which one of these activation functions would
you recommend using for the output layers
a. Sigmoid
b. Leaky ReLU
c. tanh
d. ReLU
Correct
c. Data can be stored, deleted and read from these gated cells like computer storage.
◄ CNN
Jump to...