ML Aat 2
ML Aat 2
ml aat 2.docx
Institute of Aeronautical Engineering (IARE)
Document Details
Submission ID
trn:oid:::3618:73257717 18 Pages
Download Date
File Name
ml aat 2.docx
File Size
35.8 KB
0 Missing Citation 0%
Matches that have quotation marks, but no in-text citation
Integrity Flags
0 Integrity Flags for Review
Our system's algorithms look deeply at a document for any inconsistencies that
No suspicious text manipulations found. would set it apart from a normal submission. If we notice something strange, we flag
it for you to review.
0 Missing Citation 0%
Matches that have quotation marks, but no in-text citation
Top Sources
The sources with the highest number of matches within the submission. Overlapping sources will not be displayed.
1 Publication
2 Internet
1library.net 1%
3 Internet
robots.net 1%
4 Internet
soulpageit.com 1%
5 Publication
6 Publication
7 Internet
fastercapital.com 1%
8 Internet
www.geeksforgeeks.org 1%
9 Publication
Raman Kumar, Sita Rani, Sehijpal Singh. "Machine Learning for Sustainable Manu… 1%
10 Publication
11 Internet
www.hindawi.com 1%
12 Internet
eitca.org 1%
13 Internet
hdl.handle.net 1%
14 Internet
www.ziprecruiter.com 1%
15 Publication
Peter L. Bartlett, Vitaly Maiorov, Ron Meir. "Almost Linear VC-Dimension Bounds f… 1%
16 Internet
www.coursehero.com 1%
17 Internet
ses.library.usyd.edu.au 1%
18 Internet
huggingface.co 0%
19 Publication
Antony Browne, Ron Sun. "Connectionist inference models", Neural Networks, 20… 0%
20 Internet
chatgpt-lamda.com 0%
21 Internet
slidetodoc.com 0%
22 Publication
23 Internet
www.idvip.edu.pe 0%
24 Internet
www.agu.edu.tr 0%
25 Publication
26 Internet
pergamos.lib.uoa.gr 0%
27 Publication
Vivek S. Sharma, Shubham Mahajan, Anand Nayyar, Amit Kant Pandit. "Deep Lear… 0%
28 Internet
medwinpublishers.com 0%
29 Internet
revolution.allbest.ru 0%
30 Internet
aicontentfy.com 0%
31 Internet
cibtech.org 0%
32 Publication
33 Publication
34 Publication
35 Publication
Shayan Doroudi. "The Bias-Variance Tradeoff: How Data Science Can Inform Educ… 0%
36 Internet
citeseerx.ist.psu.edu 0%
37 Internet
dokumen.pub 0%
38 Internet
ia600403.us.archive.org 0%
39 Publication
40 Publication
Jagjit Singh Dhatterwal, Kuldeep Singh Kaswan, Reenu Batra. "Nature Inspired Ro… 0%
41 Publication
42 Publication
43 Internet
bv.fapesp.br 0%
44 Internet
citizenside.com 0%
45 Internet
s3-us-west-2.amazonaws.com 0%
46 Internet
www.exploredatabase.com 0%
47 Publication
48 Publication
"Machine Learning: ECML 2001", Springer Science and Business Media LLC, 2001 0%
49 Publication
Biswadip Basu Mallik, Gunjan Mukherjee, Rahul Kar, Aryan Chaudhary. "Deep Lea… 0%
50 Publication
51 Publication
52 Publication
"Data Mining and Knowledge Discovery Handbook", Springer Science and Busine… 0%
53 Publication
Dothang Truong. "Data Science and Machine Learning for Non-Programmers - Us… 0%
54 Publication
Walker, Matthew, and Andrew Curtis. "Varying prior information in Bayesian inve… 0%
AAT – II
1. Overfitting:
14 Definition: Overfitting occurs when a machine learning model is too complex and
performs well on the training data but poorly on new, unseen data (test data or
future predictions).
9 1. Model Complexity: When a model has too many parameters relative to the
12 amount of training data, it can memorize the noise and random fluctuations in
the training data rather than learning the underlying patterns.
5 2. Noise Fitting: The model fits the random noise in the training data, which is not
representative of the broader population.
3. Insufficient Training Data: With limited training data, the model may not have
27 enough information to generalize well to new data.
4. Poor Regularization: Inadequate or absent regularization techniques (e.g., L1, L2,
Dropout) allow the model to over-emphasize minor patterns in the training data.
8 Consequence: The overfitted model becomes too specialized to the training data
and fails to generalize well to new, unseen data.
2. Underfitting:
3 Definition: Underfitting occurs when a machine learning model is too simple and
fails to capture the underlying patterns in the training data, resulting in poor
performance on both the training data and new, unseen data.
18 1. Model Simplicity: When a model is too simple or has too few parameters, it
cannot capture the complexity of the underlying relationships in the data.
2. Inadequate Features: The model may not have access to relevant features or
sufficient information to make accurate predictions.
3. Insufficient Training: The model may not have been trained for a sufficient
number of iterations or with an appropriate optimization algorithm.
4. High Bias: The model's underlying assumptions or design may be flawed, leading
to high bias and poor fit to the data.
5 Consequence: The underfitted model is too general and fails to capture the signal in
the training data, leading to poor performance on both training and new data.
44 The Ideal Balance: The goal is to find a balance between model complexity and
20 simplicity, ensuring the model generalizes well to new data while capturing the
underlying patterns in the training data. Techniques like cross-validation,
regularization, and hyperparameter tuning can help achieve this balance.
22 Here's an explanation of how high bias and low variance contribute to underfitting:
3 Underfitting: The model fails to capture the underlying patterns in the training
data, resulting in poor performance on both the training data and new, unseen
data.
1. High Bias:
2. Low Variance:
Definition: Low variance (or low statistical error) occurs when a model's
predictions are consistently similar, with minimal variation across different
training datasets (of the same size).
Contribution to Underfitting:
o Lack of adaptability: Low variance models are often too rigid, failing to
47 adapt to the nuances of the training data.
o Insensitivity to training data: The model's predictions are relatively
unchanged by the specific training data, indicating a lack of learning.
o Similar poor performance across datasets: The model will perform
similarly poorly on different training datasets, as its low variance
indicates a lack of responsiveness to the data.
Example of Low Variance:
o A model that always predicts the same value (e.g., the mean of the target
variable), regardless of the input features.
High Bias → Poor Fit → Low Variance: A model with high bias will have a poor fit
to the training data. As a result, its predictions will be consistently poor (low
variance), leading to underfitting.
Low Variance → Inability to Improve → High Bias: A model with low variance
will struggle to improve its predictions, even with more training data. This
inability to adapt is a hallmark of high bias, reinforcing the underfitting issue.
Mitigating Underfitting:
13 3. Describe the key practical difficulties in applying gradient descent rule and explain
how it is overcomed .
13 Here's a description of the key practical difficulties in applying the Gradient Descent
(GD) rule and how they are overcome:
4. Define Perceptron. State the output function. List out some of the applications of ANN.
1. Definition of Perceptron:
Input Layer: One or more input nodes that receive the input values.
Weights: Each input node is connected to the output node through a weighted
connection.
Bias: An additional input with a fixed value (usually 1) that helps the Perceptron
learn more complex patterns.
Output Layer: A single output node that generates the final output.
y = φ (Σ (wi * xi) + b)
Where:
Here are some examples of the many applications of ANNs, including Perceptrons
and more complex networks:
27 Computer Vision:
o Image Classification (e.g., object detection, facial recognition)
o Image Segmentation (e.g., self-driving cars, medical imaging)
o Object Detection (e.g., surveillance, robotics)
Natural Language Processing (NLP):
o Sentiment Analysis (e.g., text classification, opinion mining)
o Language Translation (e.g., Google Translate)
o Speech Recognition (e.g., virtual assistants, voice-controlled interfaces)
Predictive Analytics:
43 o Time Series Forecasting (e.g., stock prices, weather)
o Recommendation Systems (e.g., product suggestions, content
personalization)
o Anomaly Detection (e.g., fraud detection, network security)
Gaming and Robotics:
o Game Playing Agents (e.g., AlphaGo, poker bots)
o Robotics Control (e.g., autonomous vehicles, robotic arms)
Healthcare and Medicine:
o Disease Diagnosis (e.g., cancer detection, medical imaging analysis)
o Personalized Medicine (e.g., tailored treatment plans)
o Medical Research (e.g., drug discovery, genomics analysis)
21 What is VC Dimension?
Formal Definition:
6 Given a hypothesis space H (a set of functions) and a set of n points X = {x₁, x₂,
25 ..., xₙ } in the input space, we say that H shatters X if for every possible labeling Y =
32
25 {y₁, y₂, ..., yₙ } of the points in X, there exists a function h ∈ H such that h(xᵢ) = yᵢ for
all i = 1, 2, ..., n.
15 The VC Dimension of H, denoted by VC(H), is the largest integer n such that there
exists a set of n points X that can be shattered by H. If no such largest integer exists
51 (i.e., H can shatter arbitrarily large sets of points), then VC(H) is said to be infinite.
When dealing with infinite hypothesis spaces, the VC Dimension can be:
Finite: If the hypothesis space has a finite VC Dimension, it implies that there is a
limit to the complexity of the functions in the space. Examples include linear
classifiers in a fixed-dimensional space.
Infinite: If the hypothesis space has an infinite VC Dimension, it means that the
space can shatter arbitrarily large sets of points. This often occurs in:
o Non-parametric models (e.g., decision trees, neural networks with an
unbounded number of layers or units)
o Models with an infinite number of parameters (e.g., some types of kernel
machines)
8 Overfitting: The model may become too specialized to the training data and fail
to generalize well to new, unseen data.
5 Lack of Uniform Convergence: The model's performance on the training set may
not guarantee good performance on the test set, even with an infinite amount of
training data.
Bayesian Theory:
1. Inference: Compute the posterior probability of a node (or a set of nodes) given
evidence (observations) on other nodes.
2. Steps:
o Evidence: Observe values for some nodes.
o Propagation: Update the probabilities of all nodes based on the evidence,
using Bayes' Theorem and CPTs.
o Query: Compute the posterior probability of the node(s) of interest.
1. Simple Bayesian Network (SBN): A basic BN with a small number of nodes and
edges.
2. Dynamic Bayesian Network (DBN): Models temporal relationships between
variables.
3. Hierarchical Bayesian Network (HBN): Organizes nodes into a hierarchical
structure.
4. Hybrid Bayesian Network: Combines discrete and continuous variables.
49 CBR is a type of machine learning approach that solves new problems by adapting
1 solutions from similar, previously encountered problems (cases). In the context of
48 classification, a CBR classifier uses a database of pre-classified cases to predict the
class label of a new, unseen instance.
50 1. Case Base (Knowledge Base): A database of pre-classified cases, where each case
consists of:
o Features (Attributes): A set of relevant characteristics that describe the
case.
o Class Label (Target Variable): The corresponding class label for each case.
2. New Case (Query Case): The unseen instance to be classified, described by its
features.
10 3. Similarity Measure: A function that calculates the similarity between the new
case and cases in the case base.
1 4. Retrieval Module: Retrieves the most similar cases from the case base, based on
the similarity measure.
1 5. Adaptation Module (Optional): Adjusts the solution from the retrieved cases to
better fit the new case.
1 6. Classifier: Predicts the class label for the new case, based on the solutions from
the retrieved cases.
1. Euclidean Distance
2. Cosine Similarity
3. Jaccard Similarity
36 4. Longest Common Subsequence (LCS)
5. Dynamic Time Warping (DTW)
30 1. Case Base Quality: The accuracy of the classifier relies heavily on the quality and
representativeness of the case base.
2. Similarity Measure Selection: Choosing the most effective similarity measure can
be challenging.
3. Scalability: Can become computationally expensive with a large case base.
4. Adaptation Complexity: The adaptation module can add complexity to the
system.
2 8. Explain radial basis function. List the advantages of radial basis function network.
4 Here's an explanation of Radial Basis Function (RBF) and the advantages of Radial
Basis Function Networks (RBFNs):
between the input point and each center, where the weights are determined by a
non-linear function (typically a Gaussian or exponential function).
Where:
An RBFN is a type of neural network that uses RBFs as the activation functions in
the hidden layer. The architecture of an RBFN typically consists of:
5. Robust to Noise: RBFNs can be more robust to noisy data due to the smoothing
effect of the RBFs.
2 6. Non-Linear Mapping: RBFNs can perform non-linear mapping between the input
and output spaces.
52 7. Localized Learning: RBFNs can focus on local patterns in the data, reducing the
impact of outliers.
8. Suitable for High-Dimensional Data: RBFNs can handle high-dimensional data
with a relatively small number of hidden units.
1. Function Approximation
2. Classification
3. Regression
4. Time Series Prediction
5. Image and Signal Processing
6. Control Systems
1. Agent: * Definition: The Agent is the decision-making entity that interacts with
the Environment. * Role: The Agent observes the Environment, takes Actions, and
receives Rewards or Penalties. * Goal: The Agent's objective is to learn a Policy that
maximizes the cumulative Reward over time.
2. Environment: * Definition: The Environment is the external world that the Agent
interacts with. * Role: The Environment receives Actions from the Agent, transitions
to a new State, and provides Rewards or Penalties. * Characteristics: The
Environment can be fully or partially observable, deterministic or stochastic, and
episodic or continuous.
3. Actions (A): * Definition: Actions are the decisions made by the Agent to interact
with the Environment. * Role: Actions influence the Environment's State and the
Reward received by the Agent. * Types: Actions can be discrete (e.g., up, down, left,
right) or continuous (e.g., acceleration, steering).
4. Reward Signal (R): * Definition: The Reward Signal is a feedback mechanism that
evaluates the Agent's Actions. * Role: The Reward Signal guides the Agent to learn a
Policy that maximizes the cumulative Reward. * Characteristics: Rewards can be:
+ Positive (encouraging desired behavior) + Negative (discouraging undesired
behavior) + Sparse (provided occasionally, e.g., only when a goal is achieved)
+ Dense (provided frequently, e.g., at every time step)