CS3244 (2120) - Project Discussion 1 - Overview
CS3244 (2120) - Project Discussion 1 - Overview
3 3
Content Overview
1. The Application
2. The Model
3. The Evaluation
4 4
Designing a
Machine Learning (ML)
Application
The Machine Learning Application
6
Designing Applications
▪ SDLC
– Planning → Analysis → Design → Implementation → Maintenance → Planning → ...
7
Main Issues
1. What objectives?
– Model accuracy?
– Performance measures
▪ Quantifying the objectives
2. What data?
– Use existing dataset or collecting data?
– What do I know about the domain?
▪ Features?
▪ Hypothesis representation/space?
8
Objectives Apart from Accuracy
9
Gathering Data
10
Constructing a Good Predictor
11
Consistency with Training Data
Versus Generalisation
https://archive.ics.uci.edu/ml/datasets/Mushroom
– Variance
▪ Error from sensitivity to small fluctuations in training data
– More general hypotheses have lower variance
14 14
Overfitting & Underfitting
▪ Examples
– Decision trees
▪ Larger/deeper tree ⇒ lower bias; higher variance
– Neural Networks
▪ More hidden units ⇒ lower bias; higher variance
15 15
Simple Ideas to Improve
Generalisation Performance
16
Feature Selection
– Wrapper approach
▪ ML algorithm used to assess value of attribute sets
– Embedded approach
▪ Feature selection is part of the ML algorithm
17 17
Validation Using Cross-Validation
▪ k-Fold Cross-validation
▪ Divide training set, S, into k-folds, s1, ..., sk
For each fold si
Train model using S \ si
Test model using si
Take mean performance
▪ Wrapper-based approach
– Selecting hyperparameters
– Selecting attributes
19 19
Ensembles
▪ General idea
– Aggregate predictions of multiple hypotheses to generate an overall classification that is more accurate
▪ General motivation
– Assume k independent (i.e., uncorrelated) hypotheses
– Assume generalisation performance > 0.5
sum binomial
terms with 51+
successes
much higher
success as p
tends to 1
20 20
Ensemble Framework
Example:
Random Forest
21 21
Evaluation
22
Model Evaluation
▪ Form hypothesis test to see your model is significantly better than the
benchmark
– m × k-Fold Cross-Validation Performance
– Each value is a mean (central limit theorem applies)
– Apply t-test
23 23
Experimental Setup for Empirical Evaluation
▪ Example Walkthrough
24
Summary
25 25
Questions?
26