191IT7310Machine LearningQB
191IT7310Machine LearningQB
191IT7310Machine LearningQB
QUESTION BANK
UNIT- 1INTRODUCTION
Machine Learning – Types of Machine Learning – Supervised Learning – Unsupervised Learning –
Basic Concepts in Machine Learning – Machine Learning Process – Weight Space – Testing Machine
Learning Algorithms – A Brief Review of Probability Theory –Turning Data into Probabilities – The
Bias-Variance Trade-off.
S.NO Q&A CO K-LEVEL
PART- A
Which subfield of artificial intelligence focuses on algorithms and
models that allow computers to learn and make predictions without
explicit programming?
1 a) Machine Learning CL1
CO1.1
b) Natural Language Processing
c) Robotics
d) Computer Vision
Supervised learning is a type of machine learning where the algorithm
learns from labeled data to make predictions or decisions.
2 CO1.2 CL2
a) True
b) False
Which type of learning algorithm learns from rewards or punishments
based on its actions?
a) Supervised Learning
3 CO1.2 CL2
b) Unsupervised Learning
c) Reinforcement Learning
d) Deep Learning
Which type of learning deals with partially labeled data and combines
supervised and unsupervised learning?
a) Reinforcement Learning
4 b) Semi-Supervised Learning CO1.2 CL2
c) Transfer Learning
d) Active Learning
In supervised learning, what are the known outputs or target values
associated with the input data called?
a) Features
5 CO1.2 CL1
b) Labels
c) Training Data
d) Test Data
In which type of learning does the algorithm learn from labeled
examples?
a) Supervised Learning
6 C01.2 CL1
b) Unsupervised Learning
c) Reinforcement Learning
d) Semi-Supervised Learning
Which type of machine learning algorithm is used for continuous
target variables?
a) Classification
7 b) Regression CO1.2 CL1
c) Clustering
d) Dimensionality Reduction
Unsupervised learning is a type of machine learning where the
algorithm learns from unlabeled data to discover patterns or structures
8 in the data. CO1.3 CL2
a) True
b) False
Which technique can be used to group similar data points together
based on their characteristics?
a) Regression
9 b) Clustering CO1.3 CL2
c) Classification
d) Dimensionality Reduction
What is the process of transforming raw data into a suitable format for
machine learning algorithms called?
a) Data Collection
10 CO1.4 CL2
b) Data Preprocessing
c) Feature Engineering
d) Model Selection
Which technique is used to handle missing values in a dataset?
a) Data Imputation
11 b) Data Normalization CO1.4 CL2
c) Feature Scaling
d) Principal Component Analysis
Which step in the machine learning process involves selecting or
creating the most relevant features for the algorithm?
a) Data Collection
12 CO1.4 CL2
b) Data Preprocessing
c) Feature Selection/Engineering
d) Model Selection
What is the goal of feature engineering in machine learning?
a) Selecting the best machine learning algorithm
13 b) Transforming data into a suitable format for analysis CO1.4 CL1
c) Discovering patterns or structures in unlabeled data
d) Creating or selecting relevant features for the model
Which step of the machine learning process involves adjusting the
model's internal parameters or weights using training data?
a) Data Preprocessing
14 CO1.4 CL2
b) Feature Engineering
c) Model Selection
d) Training
Which step of the machine learning process involves applying the
trained model to new, unseen data?
a) Data Preprocessing
15 b) Model Selection CO1.4 CL2
c) Evaluation
d) Testing and Deployment
Which step in the machine learning process involves assessing the
performance of the trained model on unseen data?
a) Data Collection
16 CO1.4 CL2
b) Data Preprocessing
c) Evaluation
d) Testing and Deployment
Which term refers to the space of all possible combinations of
weights in a machine learning model?
a) Feature Space
17 CO1.5 CL1
b) Target Space
c) Weight Space
d) Prediction Space
Which type of machine learning algorithm aims to minimize the error
or loss function during training?
a) Unsupervised Learning
18 b) Reinforcement Learning CO1.6 CL2
c) Supervised Learning
d) Deep Learning
Which metric is commonly used to evaluate classification algorithms?
a) Mean Squared Error
19 b) R-squared CO1.6 CL2
c) Accuracy
d) Root Mean Squared Error
What is the main purpose of cross-validation in machine learning?
a) Selecting the best features for the model
20 b) Assessing the model's performance on unseen data CO1.6 CL1
c) Tuning the model's hyperparameters
d) Preventing overfitting of the model
Which type of machine learning algorithm aims to find the optimal
decision boundary between classes in the data?
a) Linear Regression
21 CO1.6 CL2
b) K-nearest Neighbors
c) Support Vector Machines
d) Naive Bayes
Which technique can be used to prevent overfitting in machine
learning models?
a) Regularization
22 CO1.6 CL2
b) Feature Extraction
c) Cross-validation
d) Gradient Descent
Which technique is used to reduce the dimensionality of the data
while preserving most of its important features?
a) Clustering
23 b) Principal Component Analysis CO1.6 CL1
c) Random Forests
d) Gradient Boosting
A. Bias-Variance Trade-off
B. Machine Learning Process
C. Probability Theory
D. Weight Space
d)A - 4
B-3
C-2
D-1
PART –B
In the realm of modern technological advancements, how does a
particular form of learning contribute to machines becoming
1 CO1.1 CL2
increasingly intelligent and autonomous?
Answer:
A-2
B-1
C-3
D-4
PART –B
1 Can you explain how the violation of essential assumptions in linear CO2.1 CL1
regression can impact the model's performance?
How does the use of basis function in linear regression models
2 CO2.1 CL2
enhance the model's flexibility and adaptability?
Could you provide an analysis of how the bias-variance trade-off
influences the generalization performance of machine learning
3 CO2.2 CL2
models, emphasizing the intricate equilibrium between underfitting
and overfitting?
Could you elaborate on the underlying principles of Bayesian Linear
4 Regression and shed light on how it distinguishes itself from CO2.3 CL2
classical linear regression methods, according to your understanding?
Suppose you have the following dataset representing the relationship CL3
between the number of study hours (x) and the exam scores (y):
Study Hours Exam Scores
(x) (y)
3 55
5 70
7 85
9 95
5 CO2.3
x1 x2 Class Label
2 3 0
3 4 0
5 6 1
7 8 1
6 CO2.4 CL3
9 10 1
4 5 ?
Question:
Given the dataset above, suppose we use the linear regression model.
Can you calculate the predicted class label (0 or 1) for the new data
point? What is the predicted class label for this data point based on
the given threshold 0.5?
Solve the below problem using Naïve Bayesian classifier and
find out the solution if the weather is sunny, then the player
should play or not?
7 CO2.9 CL3
Competiti
Age Type Profit
on
Old Yes Software Down
Old No Software Down
Old No Hardware Down
10 CO2.8
Mid Yes Software Down
Mid Yes Hardware Down
Mid No Hardware Up
Mid No Software Up
New Yes Software Up
New No Hardware Up
New No Software Up
State key characteristics and benefits of decision trees and random
11 forests in machine learning, and how do they differ in their CO2.9 CL1
approaches to building predictive models?
CL3
3 CO2.3 CL2
Data points: (2, 3), (3, 5), (5, 4), (7, 3), (9, 6), (10, 8)
2 CO3.2 CL3
Apply K-Means clustering algorithm to partition the data into two
clusters.
In what ways can a Dirichlet Process Mixture Model address the
3 problem of automatically determining the number of clusters in a CO3.2 CL3
dataset?
Give the limitations of spectral clustering when dealing with large
4 CO3.3 CL2
datasets and propose a possible improvement.
Provide an example where hierarchical clustering fails to produce
5 CO3.3 CL3
meaningful clusters due to its sensitivity to outliers.
Suggest strategies to mitigate the curse of dimensionality when
6 CO3.5 CL2
dealing with high-dimensional data in clustering tasks.
Dimensionality reduction can be a double-edged sword, both aiding
7 and hindering unsupervised learning tasks .give your point of view CO3.6 CL2
about this
Give a situation where PCA is inappropriate due to the underlying
8 CO3.7 CL3
structure of the data and suggest an alternative technique.
Could you provide the primary steps involved in performing Principal
9 CO3.7 CL1
Component Analysis.
Contrast the assumptions made by latent variable models with those
10 CO3.8 CL2
made by explicit clustering algorithms.
Elucidate the fundamental idea behind Latent Dirichlet Allocation and its
11 CO3.9 CL2
application in topic modelling.
Examine the challenges of LDA and propose methods to enhance the
12 CO3.9 CL2
interpretability of the results.
PART –C
In which scenarios would the EM algorithm converge slowly or fail to
1 converge for fitting a mixture model, and how can these issues be CO3.1 CL2
addressed?
i) Describe the challenges that might arise when using the K-Means
algorithm to cluster data with unevenly distributed or non-spherical
clusters, and propose potential solutions to address these challenges.
PART –B
Consider the following Bayesian network:
1 CO4.1 CL3
10 CO4.6 CL2
1 CO4.1 CL1
probability of D being true?
b) For the following Bayesian network
we know that X and Z are not guaranteed to be independent if the
value of Yis unknown. This means that, depending on the
probabilities, X and Z can beindependent or dependent if the value of
Y is unknown. Construct probabilitieswhere X and Z are independent
if the value of Y is unknown, and show thatthey are indeed
independent.
Investigate the concept of d-separation within Bayesian Networks and
its practical utility in detecting conditional independence
2 CO4.2 CL2
relationships. How does this mechanism enhance our ability to infer
probabilistic dependencies efficiently?
Provide a comprehensive illustration of the application of Markov
Random Fields in image segmentation. Detail the model's
3 CO4.3 CL3
components and step-by-step inference process to highlight its
effectiveness in capturing spatial dependencies in image data.
Delve into the challenges involved in learning the structure of
Bayesian Networks from data. Explore cutting-edge approaches and
4 CO4.4 CL2
methodologies aimed at overcoming these challenges and improving
the accuracy of model structure learning.
Elucidate the process of maximum likelihood estimation and
Bayesian parameter estimation within graphical models. Compare and
5 CO4.4 CL1
contrast their respective strengths and weaknesses in the context of
model parameter estimation.
Assess the impact of feature independence assumptions on the
performance of Naive Bayes classifiers. Propose strategies and
6 CO4.5 CL2
techniques to mitigate potential issues related to these assumptions
and enhance classifier accuracy.
characterize the concept of Markov Models and their practical
7 relevance in effectively modeling sequential data and time-series CO4.6 CL1
phenomena across diverse domains.
Engage in a detailed discussion on the key disparities between
observable and hidden states within Hidden Markov Models
8 (HMMs). Expound on the pivotal roles these states play in effectively CO4.7 CL2
modeling latent variables and observed data with real-world
applications in mind.
1. Feature Extraction
10 2. Multiple Hidden Layers
3. Sequential Ensemble Learning
4. Informative Sample Selection
Answer:
A-4
B-2
C-3
D-1
Which of the following is not a common activation function used in
neural networks?
a) Sigmoid
11
b) ReLU
c) Tanh CO5.4 CL2
d) Logistic
Which technique is used in reinforcement learning to estimate the
value of state-action pairs based on the observed rewards and the
estimated values of future states? CO5.5 CL2
12 a) Q-learning
b) Deep Q-networks (DQN)
c) Monte Carlo methods
d) Policy gradients
Which technique is used in representation learning to learn a
compressed representation of the data by encoding and decoding it
through a neural network? CO5.5 CL2
13 a) Principal Component Analysis (PCA)
b) t-SNE (t-Distributed Stochastic Neighbor Embedding)
c) Autoencoders
d) Gaussian Mixture Models (GMMs)
Which ensemble learning technique combines the predictions of
multiple models byweighting each model's prediction based on its
performance? CO5.5 CL2
14 a) Bagging
b) Boosting
c) Stacking
d) Random forest
Which technique is used in active learning to select the most
informative instances by measuring the uncertainty of the model's
predictions? CO5.5 CL2
15 a) K-means clustering
b) Decision trees
c) Support vector machines
d) Uncertainty sampling
Which technique is used in ensemble learning to train models on
different subsets of the data by resampling the data with replacement?
a) Bagging CO5.5 CL2
16
b) Boosting
c) Stacking
d) Random forest
Which ensemble learning technique is based on creating a diverse set
of models and combining their predictions using weighted averaging?
a) Bagging CO5.6 CL2
17
b) Boosting
c) Stacking
d) Random forest
Which technique is used to approximate a target function by fitting a
series of weak models one after another, each correcting the mistakes
of the previous model?
18 a) Gradient boosting
b) Deep learning CO5.6 CL2
c) Reinforcement learning
d) Support vector machines
Which technique is used in active learning to select the most
informative instances for labeling?
a) K-means clustering
19
b) Decision trees CO5.6 CL2
c) Support vector machines
d) Uncertainty sampling
Which ensemble learning technique aims to combine multiple weak
classifiers to create a strong classifier?
a) Bagging CO5.6 CL2
20
b) Boosting
c) Stacking
d) Random forest
Which technique is used to improve the generalization ability of
neural networks by artificially increasing the size of the training
dataset through creating modified copies of existing data points?
21 a) Data augmentation CO5.6 CL2
b) Dropout regularization
c) Weight decay
d) Batch normalization
Which technique is used in reinforcement learning to determine the CO5.6 CL2
value of state-action pairs based on the expected future rewards?
a) Q-learning
22
b) Deep Q-networks (DQN)
c) Monte Carlo methods
d) Policy gradients
Which technique is used in representation learning to automatically
learn feature representations from the raw input data?
a) Principal Component Analysis (PCA)
23
b) t-SNE (t-Distributed Stochastic Neighbor Embedding)
c) Autoencoders
d) Gaussian Mixture Models (GMMs) CO5.6 CL2
Which ensemble learning technique combines the predictions of
multiple models by averaging or voting to make the final prediction?
a) Bagging
24
b) Boosting CO5.6 CL2
c) Stacking
d) Random forest
Which technique is used in active learning to iteratively query the
labels for instances that are the most uncertain or difficult to classify?
a) K-means clustering
25
b) Decision trees
c) Support vector machines
d) Uncertainty sampling CO5.6 CL2
Which technique is used in ensemble learning to reduce overfitting by
training models on different subsets of the data and averaging their
predictions?
26 a) Bagging CO5.6 CL2
b) Boosting
c) Stacking
d) Random forest
Which ensemble learning technique creates a meta-model that
combines the predictions of multiple base models?
a) Bagging
27
b) Boosting
c) Stacking
d) Random forest CO5.6 CL2
Which technique is used in boosting to assign higher weights to
misclassified instances during model training?
a) Bagging
28 b) AdaBoost
c) Gradient boosting
d) Random forest CO5.6 CL2
Which technique is used in representation learning to automatically CL2
learn feature representations from the raw input data?
a) Principal Component Analysis (PCA) CO5.6
29
b) t-SNE (t-Distributed Stochastic Neighbor Embedding)
c) Autoencoders
d) Gaussian Mixture Models (GMMs)
Which ensemble learning technique combines the predictions of
multiple models by averaging or voting to make the final prediction?
a) Bagging
30 b) Boosting CO5.6 CL2
c) Stacking
d) Random forest
PART –B
delineate reinforcement learning and outline its core components that
1 CO5.1 CL2
enable agents to learn from interactions with an environment.
Could you provide an instance of a real-world application where
2 reinforcement learning is effectively utilized, showcasing its CO5.1 CL3
practicality and impact in diverse domains?
Expound on the concept of feature learning within representation
learning, elucidating how algorithms autonomously discover
3 CO5.2 CL2
meaningful features from raw data without explicit human
intervention.
Provide a succinct overview of a basic neural network's structure,
4 including its layers and their respective functions in information CO5.3 CL1
processing.
Explore the concept of active learning and highlight its advantages
5 over passive learning approaches, emphasizing the proactive role of CO5.4 CL2
the learner in selecting informative data samples.
Can you present a compelling scenario where active learning plays a
6 crucial role in significantly reducing labeling costs while achieving CO5.4 CL3
high model accuracy?
put in plain words the essence of ensemble learning and how it
7 harnesses the collective wisdom of multiple models to enhance CO5.5 CL1
overall predictive performance.
Compare and contrast bagging and boosting techniques in ensemble
8 learning, showcasing their distinctive mechanisms for aggregating CO5.5 CL2
model predictions.
Could you provide an illustrative example of an algorithm that
9 employs bagging to improve its performance, demonstrating the CO5.6 CL3
practical utility of this ensemble method?
Delve into the concept of boosting in ensemble learning, illustrating
10 its unique characteristics and how it differentiates from bagging in CO5.7 CL2
terms of model aggregation.
Discuss a compelling scenario where Gradient Boosting Machines
11 (GBMs) might be preferred over deep learning models, highlighting CO5.8 CL1
the strengths of GBMs in handling specific challenges.
Enumerate the key building blocks that constitute a deep neural
12 network, providing a concise overview of the crucial components CO5.9 CL1
responsible for its powerful learning capabilities.
PART –C
Imagine a self-driving car navigating through a complex urban
environment. The car's objective is to reach its destination while
avoiding obstacles, obeying traffic rules, and optimizing its
1 CO5.1 CL3
travel time. How reinforcement learning algorithms can
effectively balance exploration and exploitation to achieve
optimal performance in this real-world scenario.
Elaborate on strategies that can be employed to prevent or mitigate
2 CO5.2 CL2
over fitting during the training of representation learning models?
Suppose that we want to build a neural network that classifies two
dimensional data (i.e., X = [x1, x2]) into two classes: diamonds and
crosses. We have a set of training data that is plotted as follows:
3 CO5.3 CL3
4 CO5.4 CL3