CSDL Home IEEE Transactions on Pattern Analysis & Machine Intelligence 1995 vol.17 Issue No.07 - July
Issue No.07 - July (1995 vol.17)
DOI Bookmark: http://doi.ieeecomputersociety.org/10.1109/34.391407
<p><it>Abstract</it>—Inductive learning systems can be effectively used to acquire classification knowledge from examples. Many existing symbolic learning algorithms can be applied in domains with continuous attributes when integrated with a discretization algorithm to transform the continuous attributes into ordered discrete ones. In this paper, a new information theoretic discretization method optimized for supervised learning is proposed and described. This approach seeks to maximize the mutual dependence as measured by the interdependence redundancy between the discrete intervals and the class labels, and can automatically determine the most preferred number of intervals for an inductive learning application. The method has been tested in a number of inductive learning examples to show that the class-dependent discretizer can significantly improve the classification performance of many existing learning algorithms in domains containing numeric attributes.</p>
Inductive learning, classification, discretization, continuous attributes, mixed-mode attributes, maximum entropy, mutual information, uncertainty.
John Y. Ching, Andrew K. C. Wong, Keith C. C. Chan, "Class-Dependent Discretization for Inductive Learning from Continuous and Mixed-Mode Data", IEEE Transactions on Pattern Analysis & Machine Intelligence, vol.17, no. 7, pp. 641-651, July 1995, doi:10.1109/34.391407