Attribute Selection Measure
Attribute Selection Measure
information gain
Let pi be the probability that an arbitrary
tuple in D belongs to class Ci, estimated by |
Ci, D|/|D|
Expected information (entropy) needed to
Class P: buys_computer =
classify
“yes” a tuple in D:
Class N: buys_computer =
“no”
Information 9 needed
Info( D)=I( 9 ,5 )=− log
14 2 ( )− log ((after
9 5
14 14 2
5
14
)=0. 940using A to split D
attribute A
Gain(age)=Info( D)−Info ( D )=0 . 246
age
Similarly,
Gini index(CART, IBM IntelligentMiner)
Reduction
Ex. D has 9intuples
Impurity:
in buys_computer = “yes”
and 5 in “no”
The attribute provides the smallest ginisplit(D)
Suppose the attribute
(or the largest income
reduction partitions
in impurity) D into
is chosen
10 in D1the
to split : {low, medium}
node (need toand 4 in D2 all the
enumerate
possible splitting points for each attribute)
Postpruning: Remove
Assign probability branches
to each of the possible values from a “fully
grown” tree—get a sequence of progressively
Attribute construction
pruned trees
Create new attributes based on existing ones that are sparsely represented