2013 | OriginalPaper | Chapter
Decision Tree Induction: Using Frequency Tables for Attribute Selection
Activate our intelligent search to find suitable subject content or patents.
Select sections of text to find matching patents with Artificial Intelligence. powered by
Select sections of text to find additional relevant content using AI-assisted search. powered by
This chapter describes an alternative method of calculating the average entropy of the training (sub)sets resulting from splitting on an attribute, which uses frequency tables. It is shown to be equivalent to the method used in Chapter
5
but requires less computation. Two alternative attribute selection criteria, the Gini Index of Diversity and the
χ
2
statistic, are illustrated and it is shown how they can also be calculated using a frequency table.
The important issue of inductive bias is introduced. This leads to a description of a further attribute selection criterion, Gain Ratio, which was introduced as a way of overcoming the bias of the entropy minimisation method, which is undesirable for some datasets.