1.machine Learning Basics
1.machine Learning Basics
Image AI
Processing Pattern
Recognition
Symbolic Machine
Learning Learning
Computer Statistical
Robotics Vision Deep Learning
Learning
NN
Speech
NLP
Recognition
Application of AI
Learning
Learning is the process of acquiring new
understanding, knowledge, behaviors, skills, values,
attitudes, and preferences. The ability to learn is
possessed by humans, animals, and some
machines; there is also evidence for some kind of
learning in certain plants.
Source: Wikipedia
Machine Learning
Types of Learning
Databases ,Symbol and Computer
Vision
Machine Learning based on Databases. Data are required to train
machine and analysis of data using Machine Learning Algorithm.
Symbolic Learning based on symbol like human readable
character ,number and other symbol.
Computer vision tasks include methods for acquiring,
processing, analyzing and understanding digital images, and
extraction of high-dimensional data from the real world in order to
produce numerical or symbolic information, e.g. in the forms of
decisions.
What is Machine Learning?
Results Result
s
Machine Learning
Machine Learning
15
Family of
ML Models
Determine Learn by
Discover New
relationships Rewarding
Patterns
through training Actions
16
Types of Machine Learning
Type of Learning
Reinforcement
learning (RL) is an
Supervised Unsupervised
area of machine
learning is learning is a type learning concerned
the machine of machine with how software
learning task of learning algorithm agents ought to take
learning a function used to draw actions in an
that maps an input to inferences from environment in order
datasets consisting of to maximize the notion
an output based on
input data without of cumulative reward.
example input-output
What is Supervised Learning?
20
Supervised Learning
22
Un-Supervised Learning
Classification Regression
(Categorical Data) (Numerical Data)
Variable
Variable 3a
2a Variable
Variable 3b
1 Variable
Variable 4a
2b Variable
4b
Decision Trees Linear
Sok, K. (2021, December 14). Explaining Linear Regression with Hypothesis Testing and Confidence Interval. Medium.
Regression
https://khemsok97.medium.com/explaining-linear-regression-with-hypothesis-testing-and-confidence-interval-aba454e4775e
27
Unsupervised Learning Common Algorithms
Examples
Clustering
K-means Clustering
28
Machine Learning Process
Machine Learning Process
Difference between Supervised
and Unsupervised Learning
Supervised Learning Unsupervised Learning
Supervised learning algorithms are Unsupervised learning algorithms
trained using labeled data. are trained using unlabeled data.
Supervised learning needs supervision to Unsupervised learning does not need any
train the model. supervision to train the model.
Supervised learning can be used for Unsupervised learning can be used for
those cases where we know the input as those cases where we have only input
well as corresponding outputs. data and no corresponding output data.
Supervised Learning Unsupervised Learning
MACHINE LEARNING
A high-bias model will be unable to capture the dataset trend. It has a high error rate and
is considered an underfitting model. This happens because of a very simplified
algorithm. For instance, a linear regression model might be biased if the data has a
non-linear relationship .
Ways To Reduce High Bias
Since we have discussed some disadvantages of having high bias, here are some ways
to reduce high bias in machine learning.
• Use a complex model: The extremely simplified model is the main cause of high
bias. It is incapable of capturing the data complexity. In such scenarios, the model
can be made more complex.
• Increase the training data size: Increasing the training data size can help reduce
bias. This is because the model is being provided with more examples to learn
from the dataset.
• Increase the features: Increasing the number of features will increase the
complexity of the model. This improves the ability of the model to capture the
underlying data patterns.
• Reduce regularisation of the model: L1 and L2 regularisation can help
prevent overfitting and improve the model’s generalisation ability. Reducing
the regularisation or removing it completely can help improve the
performance.
What is Variance?
The variability of model prediction for a given data point which tells us the spread
of our data is called the variance of the model.
The model with high variance has a very complex fit to the training data and thus
is not able to fit accurately on the data which it hasn’t seen before.
As a result, such models perform very well on training data but have high error
rates on test data.
When a model is high on variance, it is then said to as Overfitting of Data.
Overfitting is fitting the training set accurately via complex curve and high order
hypothesis but is not the solution as the error with unseen data is high.
While training a data model variance should be kept low.
The high variance data looks as
follows.
Combination Characteristic
c
Classification
Regression
Linear Regression.
Logistic Regression.
Decision Tree.
SVM.
Naive Bayes.
kNN.
K-Means.
Random Forest.
Parametric Methods
Mean
Standard Deviation
Assumptions for Parametric
Methods
Parametric methods require several assumptions
about the data:
Normality: The data follows a normal (Gaussian)
distribution.
Homogeneity of variance: The variance of the
population is the same across all groups.
Independence: Observations are independent of
each other.
What are Parametric Methods?
Statistical Tests:
Here when we use parametric methods then the result or When we use non-parametric methods then the result or
outputs generated can be easily affected by outliers. outputs generated cannot be seriously affected by outliers.
Parametric methods have more statistical power than Non- Non-parametric methods have less statistical power than
Parametric methods. Parametric methods.
As far as the computation is considered these methods are As far as the computation is considered these methods are
computationally faster than the Non-Parametric methods. computationally slower than the Parametric methods.
Examples: Logistic Regression, Naïve Bayes Model, etc. Examples: KNN, Decision Tree Model, etc.
Thank you