Cie 1
Cie 1
1. Explain the need for Machine Learning in modern applications. Provide examples where Machine
Learning is essential.
2. Describe how Machine Learning relates to other fields such as Artificial Intelligence, Data Science,
and Statistics.
4. Discuss the major challenges in Machine Learning, including bias, overfitting, and data quality
issues.
5. Explain the Machine Learning process with a step-by-step approach, including data preprocessing,
model selection, training, and evaluation.
6. What is Big Data, and how does the Big Data Analysis Framework help in handling large-scale
datasets?
7. Explain Descriptive Statistics in Machine Learning and discuss its role in understanding data
distributions.
8. What is Univariate Data Analysis? Discuss various statistical methods and visualization techniques
used in univariate analysis.
9. Describe different data visualization techniques and their importance in Machine Learning.
Provide examples of how visual representation helps in data analysis.
10. How does Exploratory Data Analysis (EDA) contribute to better Machine Learning model
development? Discuss key techniques used in EDA.
11. Explain the differences between bivariate and multivariate data analysis. Provide examples
of statistical methods used in each type of analysis.
12. Discuss the importance of multivariate statistics in machine learning. Describe two
techniques used for analyzing multivariate data.
13. Highlight the mathematical concepts essential for understanding multivariate data. Explain
how linear algebra and calculus are applied in machine learning models.
14. Define feature engineering and discuss its significance in machine learning. Provide
examples of common techniques used to create effective features.
15. Explain the concept of dimensionality reduction. Compare and contrast Principal
Component Analysis (PCA) and t-Distributed Stochastic Neighbor Embedding (t-SNE) in
terms of methodology and applications.
16. Describe the key components involved in the design of a learning system. Discuss the
challenges faced during the development of such systems.
17. Define the concept of learning in the context of machine learning. Differentiate between
supervised, unsupervised, and reinforcement learning with suitable examples.
18. Discuss the role of modeling in machine learning. Explain the differences between
parametric and non-parametric models, providing examples of each.
19. Explain the bias-variance tradeoff in machine learning models. How does it impact the
performance and generalization of a model?
20. Discuss various evaluation metrics used to assess the performance of machine learning
models. Explain the importance of selecting appropriate metrics based on the problem
context.
21. Explain the concept of Nearest Neighbor Learning. How does it differ from model-based
learning approaches? Illustrate with an example.
22. Discuss the advantages and disadvantages of Nearest-Neighbor Learning. How does the
choice of distance metric impact its performance?
23. Describe the working mechanism of the K-Nearest-Neighbor (K-NN) algorithm. How does
increasing the value of K affect classification accuracy?
24. What are the different distance metrics used in K-Nearest-Neighbor (K-NN) classification?
Compare and contrast at least three commonly used metrics.
25. Explain the time and space complexity of the Nearest-Neighbor algorithm. How can its
efficiency be improved for large datasets?
26. What is the Weighted K-Nearest-Neighbor (WK-NN) algorithm? How does it improve
over the standard K-NN algorithm?
27. Describe different weighting techniques used in the Weighted K-Nearest-Neighbor (WK-
NN) algorithm. How does weighting influence the final classification?
28. Consider a dataset where class imbalance exists. How does the Weighted K-Nearest-
Neighbor algorithm help in handling imbalanced data? Provide an example.
29. Explain the role of hyperparameter tuning in Weighted K-Nearest-Neighbor classification.
How do parameters such as K and weight functions affect the model performance?
30. Compare and contrast the K-Nearest-Neighbor (K-NN) and Weighted K-Nearest-Neighbor
(WK-NN) algorithms in terms of accuracy, computational efficiency, and robustness to
noise.