This Article 
   
 Share 
   
 Bibliographic References 
   
 Add to: 
 
Digg
Furl
Spurl
Blink
Simpy
Google
Del.icio.us
Y!MyWeb
 
 Search 
   
34th Annual Hawaii International Conference on System Sciences ( HICSS-34)-Volume 3
Maui, Hawaii
January 03-January 06
ISBN: 0-7695-0981-9
Many decision tree (DT) induction algorithms, including the popular C4.5 family, are based on the Conditional Entropy (CE) measure family. An interesting question involves the relative performance of other entropy measure families such as Class-Attribute Mutual Information (CAMI). We therefore conducted a theoretical analysis of the CAMI family that enabled us to expose relationships with CE and correct a previous CAMI result. Our computational study showed that there was only a small variation in the performance of the two families. Since feature selection is important in DT induction, we conducted a theoretical analysis of a recently published blurring-based feature selection algorithm and developed a new feature selection algorithm. We tested this algorithm on a wider set of test problems than in the comparable study in order to identify benefits and limitations of blurring-based feature selection. These results provide theoretical and computational insight into entropy-based induction measures and feature selection algorithms.
Citation:
K. Giles, K. Bryson, Q. Weng, "Comparison of Two Families of Entropy-Based Classification Measures with and without Feature Selection," hicss, vol. 3, pp.3014, 34th Annual Hawaii International Conference on System Sciences ( HICSS-34)-Volume 3, 2001
Usage of this product signifies your acceptance of the Terms of Use.