Lec-1 Introduction
Lec-1 Introduction
Lecture-1
Topics Covered
• Introduction to Machine Learning; Decision Trees:
• Decision tree model of learning - Classification and regression using decision trees - Splitting criteria:
entropy, information gain, Gini impurity – Overfitting & Pruning in decision trees.
Introduction
• Machine Learning (ML) is considered as the most dynamic and
progressive form of human-like Artificial Intelligence.
8/13/2024 Arockiaraj S 3
Artificial Intelligence (AI) & Machine Learning (ML)?
Artificial intelligence:
• Artificial intelligence is the name
given to the process in which the
computer makes decisions,
mimicking a human.
AI ML DL
Machine learning:
• Computer makes decisions
based on experience. Computing
Growing popularity
• The word “Big Data” you keep hearing about is mainly made possible
through ML.
8/13/2024 Arockiaraj S 5
Growing popularity
Machine Learning Market Size, Share & Trends Analysis Report
8/13/2024 Arockiaraj S 6
(banking, financial services and insurance)
https://www.grandviewresearch.com/industry-analysis/machine-learning-market
8/13/2024 Arockiaraj S 7
8/13/2024 Arockiaraj S 8
Machine Learning
8/13/2024 Arockiaraj S 10
Machine Learning - Definition
• 1997, Tom Mitchell gave a definition:
Machine Learning Scientist,
Carnegie Mellon University
8/13/2024 Arockiaraj S 11
What Machine Learning does???
New Query?
8/13/2024 Arockiaraj S 12
Introduction – History
• 1950s
• Arthur Samuel (IBM)
• Program – playing Checkers game
• 1960s
• Rosenblatt
• Perceptron - Neural Network Model
• Pattern Recognition
• Later delta learning rule
• Rule for perceptron learning
• Good classifier
Introduction – History
• 1950s • 1969
• Minsky and Papert
• Arthur Samuel (IBM) • Limitation of perceptron model
• Program – playing Checkers game • Problem could not be represented
• Inseparable data distribution
• 1960s • 1970s
• Rosenblatt • Symbolic concept (AI)
• Perceptron - Neural Network Model
• Pattern Recognition • 1986
• Later delta learning rule • Quinlan – Decision Tree
• Rule for perceptron learning • ID3 Algorithm
• Good classifier • Improved: Regression
• Still popular in ML
Introduction – History
• 1990s: Machine learning involved
• 1997 – Deep Blue beats Garry Kasparov
statistics to a large extent
• 1994 – Self driving car road test • 2009 – Google builds self driving car
Labels?
• If we are trying to predict a feature based on the others, that feature is the label.
• Labeled data: Data that comes with a label. • The set of algorithms
• Unlabeled data: Data that comes without a label in which we use a
labeled dataset is
called supervised
learning.
• The spam detection model, can return only two things: spam or ham.
Thus, it is a classification model.
Regression models
• Predict numbers based on the features
• In the housing example, the features can be anything that describes a
house, such as the size, the number of rooms, the distance to the
closest school, or the crime rate in the neighborhood.
House size No. of rooms Distance to school Crime rate in the neighborhood Price
Apple:
• Size: Big
• Color: Red
• Shape: Rounded shape with a
depression at the top
Cherry:
• Size: Small
• Color: Red
• Shape: Heart-shaped to nearly globular
Banana:
• Size: Big
• Color: Green
• Shape: Long curving cylinder
Supervised Learning
Learning the physical characters of fruits
through training.
Grape:
• Size: Small
• Color: Green
• Shape: Round to oval, Bunch shape
Cylindrical
Supervised Learning
Machine already learned about the fruits through training.
Input: <Big, Red, Rounded shape>
Response: <Apple>
Input: <Small, Red, Heart-shaped>
Response: <Cherry>
Apply that
Input: <Big, Green, Long curving cylinder> knowledge to
Response: <Banana> the test data
Input: <Small, Green, Round to oval shape>
Response: <Grape>
Supervised Learning
• Already learned about the physical characters of fruits through
training.
Apple
Decision Function
/ Hypothesis
Orange
Supervised Classification
Decision Function
/ Hypothesis
Unsupervised Classification
Unsupervised Learning
• Consider physical character of that particular fruit.
• Suppose you have considered color.
• Arrange them on considering base condition as color
• Then the groups will be some thing like this.
SMALL
SMALL BIG SIZE
SIZE SIZE
BIG SIZE grapes
bananas
apples cherry
Unsupervised learning
• clustering algorithms The algorithms that group data into clusters based
on similarity.
• generative algorithms The algorithms that can generate new data points
that resemble the existing data
Clustering
• Consider the two datasets used in “Supervised learning”—the housing
dataset and the email dataset.
Clustering is an unsupervised machine learning - group the elements in our dataset into
clusters where all the data points are similar.
Clustering
• Example – email dataset
• The dataset is unlabeled, we don’t know whether each email is spam or ham.
Social
Promotions
Personal
Other applications of clustering
• Market segmentation: dividing customers into groups based on
demographics and previous purchasing behavior to create different
marketing strategies for the groups
• Hierarchical clustering: this algorithm starts by grouping the closest points together and
continuing in this fashion, until we have some well-defined groups.
• Gaussian mixture models: this algorithm does not assign a point to one cluster but
instead assigns fractions of the point to each of the existing clusters.
• For example, if there are three clusters, A, B, and C, then the algorithm could determine
that 60% of a particular point belongs to group A, 25% to group B, and 15% to group C.
Dimensionality reduction
• Simplifies data without losing too much information
• How can we reduce both the rows and the columns at the same time?
• These two algorithms express a big matrix of data into a product of smaller matrices
• a large table where each row corresponds to a user, each column to a movie, and each entry in
the matrix is the rating that the user gave the movie.
• With matrix factorization, one can extract certain features, such as type of movie, actors
appearing in the movie, and others, and be able to predict the rating that a user gives a movie,
based on these features.
• Two common types of unsupervised learning algorithms are
clustering and dimensionality reduction.
ML / DL
8/13/2024 Arockiaraj S 54
Machine Learning - Examples
• In general, to have a well-defined learning problem, we must identity these
three features:
• The class of tasks (T)
• The measure of performance (P)to be improved and
• The source of experience (E)
Wake-
up had
bath
• Reinforcement Learning 10
8/13/2024 Arockiaraj S 61
Deep Learning
“very large neural networks and huge amounts of data that we have access to”
8/13/2024 Arockiaraj S 62
Why Deep Learning?
- Performance…
8/13/2024 Arockiaraj S 63
Low level representation
8/13/2024 Arockiaraj S 64
Thank you