0% found this document useful (0 votes)
26 views2 pages

AD502 QuestionBank

Uploaded by

Atharv Tiwari
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
26 views2 pages

AD502 QuestionBank

Uploaded by

Atharv Tiwari
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 2

Prestige Institute of Engineering Management and Research, Indore

B.Tech. (Artificial Intelligence and Data Science) III Year/ V Semester

Question Bank
UNIT I: Introduction to Machine Learning
1. What is the definition of machine learning, and why is it considered significant in the context
of artificial intelligence?
2. Enumerate and explain the various stages in the machine learning life cycle.
3. Differentiate between supervised and unsupervised learning, providing examples of each.
4. Discuss three major challenges faced in the implementation of machine learning algorithms.
5. Explain the importance of data visualization in machine learning and provide an example of a
visualization technique.
6. Discuss the tradeoff between bias and variance and its implications on model performance.
7. Clarify the relationship between Artificial Intelligence (AI), Machine Learning (ML), Deep
Learning (DL), and Data Science (DS).
UNIT II: Clustering in Machine Learning
1. Define partitioning clustering and provide an example of an algorithm belonging to this
category.
2. Explain K-Means algorithm.
3. Explain the BIRCH algorithm for clustering and outline the advantages of the CURE
algorithm.
4. Discuss the concept of Gaussian Mixture Models (GMM) in clustering and explain the
Expectation-Maximization (EM) algorithm.
5. Define Maximum Likelihood Estimation (MLE) and Maximum A Posteriori (MAP) in the
context of parameter estimation.
6. Provide two real-world applications where clustering techniques are commonly used.
UNIT III: Classification Algorithm
1. Explain the logistic regression algorithm and its application in classification.
2. Outline the key principles of decision tree classification.
3. Describe the architecture of a neural network.
4. Explain the working principle of the K-Nearest Neighbors (K-NN) algorithm.
5. Discuss the theory behind Support Vector Machines (SVM) and their applications.
6. Explain the Naive Bayes algorithm, highlighting the different variants (Gaussian,
Multinomial, Bernoulli).
7. Define and explain the confusion matrix in the context of classification.
8. Discuss classification accuracy and its limitations.
9. Provide definitions and explain the significance of precision, recall, F1 score, and support in a
classification report.
UNIT IV: Ensemble Learning and Random Forest
1. Define ensemble learning and explain its purpose in improving model performance.
2. Explain the concept of voting classifiers in ensemble learning.
3. Compare Bagging and Boosting.
4. Discuss the principles of bagging and pasting in the context of ensemble methods.
5. Define out-of-bag evaluation and explain its significance in ensemble learning.
6. Discuss the concepts of random patches and random subspaces in building diverse models.
7. Describe the structure and functioning of Random Forests.
8. Discuss the concept of Extra-Trees and the importance of feature importance in Random
Forests.
9. Explain the boosting technique in ensemble learning.
10. Discuss the AdaBoost and Gradient Boosting algorithms.
UNIT V: Dimensionality Reduction
1. Define the curse of dimensionality and explain its impact on machine learning models.
2. Explain the main approaches for dimensionality reduction, including projection and manifold
learning.
3. Discuss Principal Component Analysis (PCA) and its objective in preserving variance.
4. Explain the process of projecting data down to a lower-dimensional space.
5. Discuss the importance of explained variance ratio in PCA.
6. Discuss methods for choosing the right number of dimensions in PCA.
7. Explain how PCA can be used for data compression.
8. Define PAC (Probably Approximately Correct) learning model.
9. Explain the VC (Vapnik-Chervonenkis) dimension and its relevance in machine learning
theory.

You might also like