ObjectiveQ&a Mid-I NNDL
ObjectiveQ&a Mid-I NNDL
2. Which layer of a feedforward neural network connects directly to the input data?
A) Input layer
B) Hidden layer
C) Output layer
D) Bias layer
Answer: A) Input layer
3. Which activation function is commonly used for binary classification problems in ANNs?
A) Sigmoid
B) ReLU
C) Tanh
D) Softmax
Answer: A) Sigmoid
5. Which type of neural network is specifically designed for sequence data, such as time
series or natural language processing?
A) Convolutional Neural Network (CNN)
B) Recurrent Neural Network (RNN)
C) Feedforward Neural Network (FNN)
D) Radial Basis Function Network (RBFN)
Answer: B) Recurrent Neural Network (RNN)
6. In a convolutional neural network (CNN), what is the primary operation performed by
convolutional layers?
A) Matrix multiplication
B) Pooling
C) Weight normalization
D) Element-wise multiplication
Answer: A) Matrix multiplication
7. Which type of neural network architecture is often used for image classification tasks?
A) LSTM
B) GRU
C) CNN
D) Autoencoder
Answer: C) CNN
9. In a Long Short-Term Memory (LSTM) network, what is the role of the forget gate?
A) Deciding which information to discard from the cell state
B) Determining the output of the current time step
C) Storing long-term information
D) Managing the input gate
Answer: A) Deciding which information to discard from the cell state
10. Which of the following is NOT a commonly used loss function in neural network
training?
A) Mean Squared Error (MSE)
B) Cross-Entropy Loss
C) Hinge Loss
D) K-Means Loss
Answer: D) K-Means Loss
11. What does "backpropagation" refer to in the context of neural networks?
A) The process of propagating gradients backward through the network to update weights.
B) The initial forward pass of data through the network.
C) A technique for selecting the most important features in the input data.
D) The activation function used in the output layer.
Answer: A) The process of propagating gradients backward through the network to update
weights.
14. In the context of training a neural network, what does "overfitting" mean?
A) The network is not able to fit the training data.
B) The network is too simple and cannot learn from the data.
C) The network has learned the training data too well and does not generalize to new data.
D) The network converges too quickly during training.
Answer: C) The network has learned the training data too well and does not generalize to
new data.
15. What does "epoch" refer to during the training of a neural network?
A) The number of neurons in the input layer.
B) The number of layers in the network.
C) One complete pass through the entire training dataset during training.
D) The learning rate used for updating weights.
Answer: C) One complete pass through the entire training dataset during training.
19. What is the typical approach for training a supervised learning network?
A) Providing input data without labels
B) Providing input data with labels, allowing the network to learn the mapping from input
to output
C) Providing only output labels without input data
D) Providing input data and asking the network to generate labels
Answer: B) Providing input data with labels, allowing the network to learn the mapping
from input to output
24. In the context of neural networks, what is the primary purpose of the backpropagation
algorithm?
25. During the backpropagation process, what is typically used to adjust the model's weights?
27. Which type of unsupervised learning algorithm is commonly used for dimensionality
reduction?
A) K-Means Clustering
B) Principal Component Analysis (PCA)
C) Recurrent Neural Network (RNN)
D) Gradient Descent
29. Which unsupervised learning method aims to find dense regions of data points in a high-
dimensional space?
A) Hierarchical Clustering
B) Gaussian Mixture Model (GMM)
C) Density-Based Spatial Clustering of Applications with Noise (DBSCAN)
D) Apriori Algorithm
A) Image classification
B) Dimensionality reduction and feature learning
C) Natural language processing (NLP)
D) Reinforcement learning
31. Which type of unsupervised learning model is capable of generating new, similar data
points after training?
33. What is the main objective of Fixed Weight Competitive Networks (WTA)?
Answer: B) Only one neuron, the winner, produces an output of 1, while all others output 0
35. What is the typical activation function used in Fixed Weight Competitive Networks?
A) Sigmoid
B) ReLU (Rectified Linear Unit)
C) Linear
D) Step function
36. In Fixed Weight Competitive Networks, what is the role of lateral inhibition?
Answer: C) It suppresses the activity of neighboring neurons when one neuron wins
A) Image classification
B) Clustering and pattern recognition
C) Natural language processing (NLP)
D) Reinforcement learning
39. How do Maxnet Hamming Networks update their neuron activations during training?
Answer: C) By comparing each neuron's output with a threshold and selecting the winner
41. In a Maxnet Hamming Network, what does the winner neuron represent?
Answer: D) The cluster or group to which the input data point belongs
45. How are codebook vectors updated during the training process in Learning Vector
Quantization?
Answer: C) They are adjusted based on the input data and the winner-takes-all principle
47. What are the two main layers in a Counter Propagation Network?
48. In Counter Propagation Networks, what is the role of the codebook layer?
49. What is the primary goal of Adaptive Resonance Theory (ART) networks?
Answer: C) It defines the level of similarity between input patterns and existing clusters
1. In a basic artificial neural network (ANN), the primary building block is the ________.
Answer: Neuron
2. The weights in a neuron represent the ________ of the connections.
Answer: Strength
3. The function applied to the weighted sum of inputs in a neuron is called the ________
function.
Answer: Activation
4. In supervised learning networks, the role of the supervisor is to provide ________ data for
training.
Answer: Labeled
5. The primary objective of a supervised learning network is to map input data to ________
outputs.
Answer: Target
6. The process of propagating gradients backward through the network to update weights is
known as ________.
Answer: Backpropagation
7. Perceptron networks are typically used for ________ classification tasks by finding linear
decision boundaries.
Answer: Binary
8. The step function in a perceptron helps it make ________ decisions.
Answer: Binary
9. Adaptive Linear Neuron (Adaline) networks use a ________ function for continuous
output.
Answer: Linear
10. The learning rule in Adaline networks is based on minimizing the ________ between
predicted and target values.
Answer: Mean Squared Error (MSE)
11. The process of fine-tuning neural network weights to minimize errors is the primary task
of ________ networks.
Answer: Back-propagation
12. Associative Memory Networks are designed for ________ recall of patterns from partial
or noisy input.
Answer: Pattern
13. Training algorithms for pattern association, such as Hebbian learning and delta rule, are
used to store and retrieve ________ in associative memory networks.
Answer: Patterns
14. The Bidirectional Associative Memory (BAM) network can store and retrieve patterns in
a ________ manner.
Answer: Bidirectional
15. The Hopfield network is a type of recurrent neural network used for ________.
Answer: Auto-associative memory
16. In a Hopfield network, patterns are stored as stable ________.
Answer: Equilibrium points
17. The training algorithm for BAM involves updating the weights based on the ________
product of patterns.
Answer: Outer
18. In Hopfield networks, the diagonal elements of the weight matrix are typically set to
________.
Answer: Zero
19. In a basic feedforward neural network, the process of propagating data from the input
layer to the output layer is called the ________ pass.
Answer: Forward
20. The activation function of a neuron introduces ________ into the network, allowing it to
learn complex patterns.
Answer: Non-linearity
21. A loss or cost function is used in supervised learning to measure the network's accuracy
on the ________ data.
Answer: Training
22. Dropout layers are used in neural networks to reduce ________, which can occur during
training.
Answer: Overfitting
23. In Convolutional Neural Networks (CNNs), convolutional layers primarily perform
________ operations.
Answer: Matrix multiplication
24. Long Short-Term Memory (LSTM) networks are designed for handling ________ data,
such as time series or natural language.
Answer: Sequence
25. In the context of backpropagation, the learning rate and ________ information are
typically used to adjust model weights.
Answer: Gradient
26. Unsupervised learning is a type of machine learning where the model learns from
________ data without explicit labels.
Answer: Unlabeled
27. Fixed Weight Competitive Networks, also known as Winner-Takes-All networks, use
________ weights for the connections between neurons.
Answer: Fixed
28. In Maxnet networks, during the competition phase, only one neuron, the ________,
produces an output of 1, while all others output 0.
Answer: Winner
29. Hamming Networks are often used for ________ and pattern recognition tasks.
Answer: Clustering
30. Kohonen Self-Organizing Feature Maps (SOM) are a type of unsupervised learning
network used for ________.
Answer: Dimensionality reduction
31. In Learning Vector Quantization (LVQ), codebook vectors are used to represent
________ or clusters in data.
Answer: Prototypes
32. In Counter Propagation Networks (CPN), codebook vectors are adjusted based on the
input data and the ________ principle.
Answer: Winner-takes-all
33. The primary role of Adaptive Resonance Theory (ART) networks is to discover
________ in data.
Answer: Clusters
34. The vigilance parameter in ART networks defines the level of ________ between input
patterns and existing clusters.
Answer: Similarity
35. Special Networks encompass various types of neural networks designed for specific tasks
or applications, such as image recognition or ________.
Answer: Natural language processing (NLP)
36. The Hopfield network is a type of recurrent neural network used for ________.
Answer: Auto-associative memory
37. The Boltzmann machine is a type of network that models the probability distribution of
________.
Answer: Binary variables
38. In Restricted Boltzmann Machines (RBMs), visible and hidden units have ________
connections between them.
Answer: Undirected
39. Convolutional Neural Networks (CNNs) are designed for tasks like image recognition
and utilize ________ layers for feature extraction.
Answer: Convolutional
40. In a Radial Basis Function Network (RBFN), radial basis functions are used as ________
functions.
Answer: Activation
41. The main purpose of Recurrent Neural Networks (RNNs) is to handle ________ data,
such as time series.
Answer: Sequential
42. The primary goal of Long Short-Term Memory (LSTM) networks is to capture ________
dependencies in data.
Answer: Long-term
43. In Echo State Networks (ESNs), the ________ is a reservoir of randomly connected
recurrent neurons.
Answer: Reservoir
44. The self-organizing property of Self-Organizing Maps (SOM) is used to arrange neurons
in a ________.
Answer: Topological map
45. In a Hopfield network, patterns are stored as stable ________ points.
Answer: Equilibrium
46. In Adaptive Resonance Theory (ART) networks, a category is formed when the input
pattern resonates with a stored ________.
Answer: Prototype
47. Growing Neural Gas (GNG) is an unsupervised learning network used for ________.
Answer: Clustering and data visualization
48. Liquid State Machines (LSMs) utilize a large reservoir of ________ connected neurons
for computation.
Answer: Randomly
49. In Extreme Learning Machines (ELMs), the weights connecting the hidden layer to the
output layer are initialized ________.
Answer: Randomly
50. A Denoising Autoencoder is designed to remove noise or corruption from ________ data.
Answer: Input