Machine Learning One Mark Answers
Machine Learning One Mark Answers
A: Unsupervised learning is a type of machine learning where the model is trained on data
without labeled responses. The algorithm tries to identify hidden patterns or intrinsic
structures in the input data, commonly using techniques like clustering (e.g., K-Means,
DBSCAN) and dimensionality reduction (e.g., PCA).
Q: What is Regression?
A: A confidence interval is a range of values, derived from a dataset, that is likely to contain
the value of an unknown population parameter. It is associated with a confidence level (e.g.,
95%), indicating the probability that the interval includes the true parameter.
Q: How to Estimate Binomial Proportions?
A: Binomial proportions can be estimated by dividing the number of successes (x) by the
total number of trials (n), i.e., p̂ = x/n. To estimate with a confidence interval, one can use
methods like the Wald interval or Wilson score interval.
Q: When to Use Naïve Bayes Classifier?
A: Naïve Bayes is most effective when the input features are conditionally independent
given the class label. It is widely used in text classification, spam filtering, sentiment
analysis, and recommendation systems due to its simplicity and efficiency.
Q: What is Maximum Likelihood Estimation (MLE)?
A: 1. Supervised Learning
2. Unsupervised Learning
3. Semi-Supervised Learning
4. Reinforcement Learning
5. Self-Supervised Learning
Q: In What Way Boosting is Different from Bagging?
A: Bagging trains multiple models in parallel using random subsets and aggregates their
results (e.g., Random Forest), while Boosting trains models sequentially, where each new
model focuses on correcting the previous model’s errors (e.g., AdaBoost). Boosting reduces
bias, while bagging reduces variance.
A: Supervised learning involves training a model on a labeled dataset, where the input data
is paired with the correct output. It is used for tasks like classification and regression.
Q: Define Variance.
A: Variance in machine learning refers to the amount by which the model’s predictions
would change if it were trained on a different dataset. High variance often indicates
overfitting.
Q: What is Confidence Interval?
A: Feature extraction transforms raw data into meaningful features to improve learning
accuracy and efficiency. Examples include PCA and LDA.
Q: What is Decision Tree Learning?
A: Decision Tree Learning uses a tree structure where nodes represent tests on attributes
and leaves represent decisions or outputs. It’s used for classification and regression.
A: 1. Supervised Learning
2. Unsupervised Learning
3. Reinforcement Learning
Q: Explain About Confidence Interval.
A: Choose k using cross-validation to balance bias and variance. Small k may lead to
overfitting; large k may oversimplify. Odd values are often preferred for classification.
Q: Explain the Use of SVM.
A: Support Vector Machines (SVMs) are supervised models used for classification and
regression. They work by finding the optimal hyperplane that separates data classes with
maximum margin.
Q: Explain the Need for Ensemble Methods.
A: DBSCAN is used because it can detect clusters of arbitrary shapes, does not require the
number of clusters in advance, and identifies noise points (outliers).