The Community for Technology Leaders
RSS Icon
Subscribe
Issue No.07 - July (1995 vol.17)
pp: 641-651
ABSTRACT
<p><it>Abstract</it>—Inductive learning systems can be effectively used to acquire classification knowledge from examples. Many existing symbolic learning algorithms can be applied in domains with continuous attributes when integrated with a discretization algorithm to transform the continuous attributes into ordered discrete ones. In this paper, a new information theoretic discretization method optimized for supervised learning is proposed and described. This approach seeks to maximize the mutual dependence as measured by the interdependence redundancy between the discrete intervals and the class labels, and can automatically determine the most preferred number of intervals for an inductive learning application. The method has been tested in a number of inductive learning examples to show that the class-dependent discretizer can significantly improve the classification performance of many existing learning algorithms in domains containing numeric attributes.</p>
INDEX TERMS
Inductive learning, classification, discretization, continuous attributes, mixed-mode attributes, maximum entropy, mutual information, uncertainty.
CITATION
John Y. Ching, Andrew K. C. Wong, Keith C. C. Chan, "Class-Dependent Discretization for Inductive Learning from Continuous and Mixed-Mode Data", IEEE Transactions on Pattern Analysis & Machine Intelligence, vol.17, no. 7, pp. 641-651, July 1995, doi:10.1109/34.391407
24 ms
(Ver 2.0)

Marketing Automation Platform Marketing Automation Tool