Lecture 01 - Introduction To ML - 4p
Lecture 01 - Introduction To ML - 4p
CSE655
Deep Learning
Introduction
Spring 2025
Introduction to Machine Learning Most of the following slides are adapted E. Alpaydin
© 2016-2025 Yakup Genc & Y. Sinan Akgul
1 2
• Most of the real-world problems are: • … a computer program that can learn from experience
– NP-Hard with respect to some class of tasks and performance
measure …” – Mitchell 1997
• scene matching, big data problems …
• “ … ML is the science of getting computers to learn,
– Ill-defined without explicitly programmed…” –Ng
• 3D reconstruction from a single image, missing data
• ML is more than just memorizing facts:
– The right answer is subjective – learning the underlying structure of the problem or data
• Image segmentation, language, …
• Also known as:
– Hard to model – Regression
• scene classification, customer behavior
– Pattern recognition
• Machine Learning tries to use statistical reasoning to find approximate – Data mining
solutions for tackling the above difficulties – Automated programming
Joshi et al Dönnes
March 2025 Deep Learning 3 March 2025 Deep Learning 4
3 4
1
3/11/2025
Test Data
# of experiences observed outcome observed input
Training
Data Magic Black Box
Model
(Learning Machine)
Training data: Sample images of the class of objects to be learned hypothesis/program space performance measure selected hypothesis
Model: The model can identify any koala images...
Dönnes
March 2025 Deep Learning 5 March 2025 Deep Learning 6
5 6
Joshi et al
March 2025 Deep Learning 7 March 2025 Deep Learning 8
7 8
2
3/11/2025
9 10
11 12
3
3/11/2025
13 14
15 16
4
3/11/2025
17 18
α1
19 20
5
3/11/2025
21 22
23 24
6
3/11/2025
Input Encoding
25 26
Predictions TP
Test Data tp fp
(true positive) (false positive)
FP predicted class Correct result Unexpected result
+1 (expectation) fn tn
Model (false negative) (true negative)
TN Missing result Correct absence of result
-1
Precision = Accuracy =
= +1 FN
= -1 Recall =
27 28
7
3/11/2025
tp fp 1
(true positive) (false positive) 𝐶
predicted class Correct result Unexpected result 𝐶 𝐶
(expectation) fn tn
(false negative) (true negative) 𝐶
Missing result Correct absence of result true positive rate
𝐶
29 30
31 32
8
3/11/2025
• Assumes equal cost for both kinds of errors • 80% accuracy = 20% error
– cost(b-type-error) = cost (c-type-error) • suppose learning increases accuracy from 80% to 90%
• Is 99% accuracy good? • error reduced from 20% to 10%
– can be excellent, good, mediocre, poor, terrible • 50% reduction in error
– depends on problem
• 99.90% to 99.99% = 90% reduction in error
• is 10% accuracy bad?
• 50% to 75% = 50% reduction in error
– information retrieval
• can be applied to many other measures
• BaseRate = accuracy of predicting predominant class (on most problems
obtaining BaseRate accuracy is easy)
33 34
Learning Theories •
space complexity)
Easier to explain
(more interpretable)
• Generalizes better (lower
variance - Occam’s razor)
35 36
9
3/11/2025
Model Selection & Generalization Bias and Variance Tradeoff (Simple Problem)
37 38
Bias Problem
– The hypothesis space made available by a particular classification method does
not include sufficient hypotheses
Variance Problem
– The hypothesis space made available is too large for the training data, and the
selected hypothesis may not be accurate on unseen data
39 40
10
3/11/2025
• Imagine that the training set was drawn at random from a whole set of
training sets
• The squared loss can be decomposed into a “bias” term and a
“variance” term:
– Bias = systematic error in the model’s estimates
– Variance = noise in the estimates cause by sampling noise in the training set
41 42
11