Module 2
Module 2
Supervised
Learning
Copyright © 2018 McGraw Hill Education, All Rights Reserved.
PROPRIETARY MATERIAL © 2018 The McGraw Hill Education, Inc. All rights reserved. No part of this PowerPoint slide may be displayed, reproduced or distributed in any form or by any
means, without the prior written permission of the publisher, or used beyond the limited distribution to teachers and educators permitted by McGraw Hill for their individual course preparation.
If you are a student using this PowerPoint slide, you are using it without permission.
Inductive learning
Task of inductive learning:
• Given a collection of examples of a function , returns
a function that approximates .
• The approximating function is called hypothesis
function.
• The unknown true function correctly maps the input
space (of the entire data) to the output space Y.
• Central aim of designing is to suggest decisions for
unseen patterns.
• Better approximation of leads to better generalization.
• Generalization performance is the fundamental problem
in inductive learning.
• Off-training set error—the error on points not in the
training set, is used as a measure of generalization
performance.
• Inductive learning assumes that the best hypothesis
regarding unseen patterns is the one induced by the
observed training set
Occam’s Razor Principle
• A simpler algorithm can be expected to perform better
on a test set.
• “simpler” – may stand for fewer parameters, lesser
training time, fewer features and so forth.
• Generally searching is stopped for a design when the
solution is “good enough” and not the optimal one.
• Occam’s razor principle recommends hypothesis
functions that avoid overfitting of the training data.
Overfitting
• With increase in complexity, the training set’s
performance increases but performance of the test set
decreases, then overfitting has happened.
• FP = FN = 0 is desired.
rarely miss abnormal event.
True Negative Rate