11 Decision Trees
11 Decision Trees
Dr Subharag Sarkar
Content
Introduction
Decision Trees
Choosing an Attribute
Entropy
Information Gain
Summary
Introduction
Decision tree induction is one of the simplest and yet most successful
forms of machine learning.
We first describe the representation—the hypothesis space—and then
show how to learn a good hypothesis.
A decision tree represents a function that takes as input a vector of
attribute values and returns a “decision”—a single output value.
The input and output values can be discrete or continuous.
For now, we will concentrate on problems where the inputs have
discrete values and the output has exactly two possible values; this is
Boolean classification, where each example input will be classified as
true (a positive example) or false (a negative example).
Decision Trees
Here we see two different attributes chosen at the root of the decision
tree.
At each node, green stands for a training example that waited, red stands
for a training example that did not wait.
Which attribute seems better (more useful) to you?
The Patrons attribute is more useful because it separates better the green
from the reds.
Choosing an Attribute