Issue No. 11 - Nov. (2012 vol. 24)
DOI Bookmark: http://doi.ieeecomputersociety.org/10.1109/TKDE.2011.149
Qinghua Hu , Harbin Institute of Technology, Harbin
Xunjian Che , Harbin Institute of Technology, Harbin
Lei Zhang , The Hong Kong Polytechnic University, Hong Kong
David Zhang , The Hong Kong Polytechnic University, Hong Kong
Maozu Guo , Harbin Institute of Technology, Harbin
Daren Yu , Harbin Institute of Technology, Harbin
In many decision making tasks, values of features and decision are ordinal. Moreover, there is a monotonic constraint that the objects with better feature values should not be assigned to a worse decision class. Such problems are called ordinal classification with monotonicity constraint. Some learning algorithms have been developed to handle this kind of tasks in recent years. However, experiments show that these algorithms are sensitive to noisy samples and do not work well in real-world applications. In this work, we introduce a new measure of feature quality, called rank mutual information (RMI), which combines the advantage of robustness of Shannon's entropy with the ability of dominance rough sets in extracting ordinal structures from monotonic data sets. Then, we design a decision tree algorithm (REMT) based on rank mutual information. The theoretic and experimental analysis shows that the proposed algorithm can get monotonically consistent decision trees, if training samples are monotonically consistent. Its performance is still good when data are contaminated with noise.
Entropy, Decision trees, Mutual information, Rough sets, Algorithm design and analysis, Robustness, Noise measurement, decision tree, Monotonic classification, rank entropy, rank mutual information
D. Zhang, L. Zhang, X. Che, Q. Hu, M. Guo and D. Yu, "Rank Entropy-Based Decision Trees for Monotonic Classification," in IEEE Transactions on Knowledge & Data Engineering, vol. 24, no. , pp. 2052-2064, 2012.