0% found this document useful (0 votes)
14 views

Introduction To Machine Learning

Basic concepts of ML

Uploaded by

Ryuji Channeru
Copyright
© © All Rights Reserved
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
14 views

Introduction To Machine Learning

Basic concepts of ML

Uploaded by

Ryuji Channeru
Copyright
© © All Rights Reserved
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
You are on page 1/ 10

Introduction to

Machine Learning
Machine learning (ML) is a field of artificial intelligence (AI) that
focuses on the development of computer systems that can learn
from data without explicit programming. It enables machines to
make predictions, classifications, and decisions based on
patterns identified in data. ML algorithms are designed to
improve their performance over time by learning from new data
and experiences. Machine learning is transforming various
industries, including healthcare, finance, transportation, and
more.
Supervised Learning Models
1 Definition 2 Applications 3 Types of Supervised Learning
Supervised learning involves Supervised learning is widely Common types of supervised
training a model on labeled data, used for tasks such as image learning include regression
where each data point has a classification, spam detection, (predicting continuous values)
corresponding output or target fraud detection, and sentiment and classification (predicting
value. The model learns the analysis. By learning from labeled categorical labels). Regression
relationship between the input data, models can effectively models are used for tasks such as
features and the output labels, classify objects, detect predicting house prices or stock
enabling it to make predictions anomalies, and predict future prices, while classification models
on unseen data. events. are used for tasks such as
identifying spam emails or
classifying images.
Unsupervised Learning Models
Definition Applications Types of Unsupervised
Learning
Unsupervised learning focuses on Unsupervised learning is used in
discovering patterns and various applications, including Common types of unsupervised
structures in unlabeled data. The customer segmentation, anomaly learning include clustering,
model is not provided with any detection, and dimensionality association rule learning, and
target values and must learn from reduction. By uncovering hidden dimensionality reduction.
the data itself to identify patterns in data, models can Clustering algorithms group data
relationships and insights. segment customers into groups points into clusters based on
with similar characteristics, similarity, while association rule
identify unusual events, and learning identifies relationships
simplify complex data sets. between different items in a
dataset. Dimensionality reduction
techniques reduce the number of
variables in a data set while
preserving important information.
Reinforcement Learning Models
Definition
Reinforcement learning involves training an agent to interact with an
environment and learn from its actions. The agent receives rewards or penalties
for its actions, which guide it towards maximizing its cumulative reward over
time.
Applications
Reinforcement learning is used in various applications, including game playing,
robotics, and control systems. It enables machines to learn complex behaviors,
such as playing games at a superhuman level or controlling robots in dynamic
environments.
Types of Reinforcement Learning
Common types of reinforcement learning include Q-learning, SARSA, and deep
reinforcement learning. These algorithms differ in how they learn and represent
the environment, but they all share the goal of maximizing reward through
interaction and learning.
Linear Regression
Definition Applications
Linear regression is a supervised Linear regression is commonly used for
learning model used to predict tasks such as predicting house prices,
continuous target variables. It assumes a stock prices, or sales revenue. It can also
linear relationship between the input be used for forecasting time series data,
features and the output variable, and the such as predicting future demand or
model learns a linear equation to sales.
represent this relationship.

Assumptions Limitations
Linear regression models make several Linear regression models are not
assumptions about the data, including suitable for predicting non-linear
linearity, normality of residuals, and relationships. They can also be sensitive
homoscedasticity. It's important to to outliers, and they may not perform
validate these assumptions before using well when the data contains a high
a linear regression model. degree of multicollinearity.
Logistic Regression
Definition Logistic regression is a supervised learning
model used to predict categorical target
variables. It uses a sigmoid function to
transform the linear combination of input
features into a probability between 0 and 1,
which represents the likelihood of the
target variable belonging to a particular
class.
Applications Logistic regression is commonly used for
tasks such as spam detection, fraud
detection, and sentiment analysis. It can
also be used for predicting customer churn
or credit risk.
Assumptions Logistic regression models make similar
assumptions to linear regression, including
linearity, normality of residuals, and
homoscedasticity. However, they also
assume that the data is linearly separable,
meaning that the classes can be separated
by a linear boundary.

Limitations Logistic regression models are not suitable


for predicting non-linear relationships. They
can also be sensitive to outliers, and they
may not perform well when the data
contains a high degree of multicollinearity.
Decision Trees
1 Definition
Decision trees are supervised learning models that use a tree-like structure to
represent a series of decisions and their corresponding outcomes. They learn from
the data to create a hierarchical structure that splits the data based on specific
features, ultimately leading to a prediction.

2 Applications
Decision trees are widely used in various applications, including customer
segmentation, risk assessment, and medical diagnosis. They can be used for both
classification and regression tasks, and their interpretability makes them valuable
for understanding the decision-making process.

3 Types of Decision Trees


There are several types of decision trees, including ID3, C4.5, and CART. These
algorithms differ in how they select features for splitting and how they handle
missing values. The choice of algorithm depends on the specific data set and the
desired outcome.

4 Limitations
Decision trees can be prone to overfitting, especially when the data is noisy or has a
high number of features. They can also be sensitive to changes in the data, which
can lead to instability in the model. However, techniques like pruning and bagging
can help mitigate these limitations.
Random Forests

Ensemble Learning Averaging Predictions


Random forests are an ensemble learning The final prediction is made by averaging
method that combines multiple decision the predictions of all the individual trees in
trees to improve prediction accuracy and the forest. This averaging process helps to
reduce overfitting. Each tree is trained on a reduce the variance of the predictions and
random subset of the data and features, improve the overall accuracy of the model.
creating a diverse set of models.

Applications Advantages
Random forests are widely used in various Random forests offer several advantages
applications, including image classification, over single decision trees, including
object detection, and medical diagnosis. improved accuracy, reduced overfitting, and
Their robustness and accuracy make them a increased robustness to noise and outliers in
powerful tool for tackling complex machine the data.
learning problems.
Support Vector Machines

Definition Kernel Trick


Support vector machines (SVMs) are supervised learning SVMs employ the kernel trick to handle non-linearly
models that find an optimal hyperplane that separates separable data. By transforming the data into a higher-
different classes in the data. They aim to maximize the dimensional space, SVMs can find a linear separation in
margin between the hyperplane and the closest data this transformed space, effectively separating classes that
points from each class, known as support vectors. were not linearly separable in the original space.
K-Nearest Neighbors
K-Nearest Neighbors (KNN) is a simple, yet effective supervised learning algorithm used for both classification
and regression tasks. It relies on the idea of finding the "k" nearest data points to a new data point and making
predictions based on their labels.

KNN classifies a new data point by assigning it the class that is most prevalent among its nearest neighbors. For
regression tasks, it predicts the value of a new data point by averaging the values of its nearest neighbors.

The choice of the "k" value and the distance metric used to calculate nearest neighbors can significantly impact
the performance of KNN.

You might also like