This Article 
 Bibliographic References 
 Add to: 
Class Noise Handling for Effective Cost-Sensitive Learning by Cost-Guided Iterative Classification Filtering
October 2006 (vol. 18 no. 10)
pp. 1435-1440
Recent research in machine learning, data mining, and related areas has produced a wide variety of algorithms for cost-sensitive (CS) classification, where instead of maximizing the classification accuracy, minimizing the misclassification cost becomes the objective. These methods often assume that their input is quality data without conflict or erroneous values, or the noise impact is trivial, which is seldom the case in real-world environments. In this paper, we propose a Cost-guided Iterative Classification Filter (CICF) to identify noise for effective CS learning. Instead of putting equal weights on handling noise in all classes in existing efforts, CICF puts more emphasis on expensive classes, which makes it attractive in dealing with data sets with a large cost-ratio. Experimental results and comparative studies indicate that the existence of noise may seriously corrupt the performance of the underlying CS learners and by adopting the proposed CICF algorithm, we can significantly reduce the misclassification cost of a CS classifier in noisy environments.

[1] M. Tan, “Cost-Sensitive Learning of Classification Knowledge and Its Applications in Robotics,” Machine Learning, vol. 13, pp. 7-33, 1993.
[2] P. Turney, “Cost-Sensitive Classification: Empirical Evaluation of a Hybrid Genetic Decision Tree Induction Algorithm,” J. AI Research, vol. 2, pp. 369-409, 1995.
[3] M. Pazzani, C. Merz, P. Murphy, K. Ali, T. Hume, and C. Brunk, “Reducing Misclassification Costs,” Proc. 11th Int'l Conf. Machine Learning, 1994.
[4] P. Domingos, “MetaCost: A General Method for Making Classifiers Cost Sensitive,” Proc. Fifth Int'l Conf. Knowledge Discovery and Data Mining, 1999.
[5] B. Zadrozny, J. Langford, and N. Abe, “Cost-Sensitive Learning by Cost-Proportionate Example Weighting,” Proc. Third Int'l Conf. Data Mining, 2003.
[6] C. Brodley and M. Friedl, “Identifying Mislabeled Training Data,” J. AI Research, vol. 11, pp. 131-167, 1999.
[7] X. Zhu, X. Wu, and Q. Chen, “Eliminating Class Noise in Large Data Sets,” Proc. 20th Int'l Conf. Machine Learning, 2003.
[8] X. Zhu, X. Wu, and Q. Chen, “Bridging Local and Global Data Cleansing: Identifying Class Noise in Large, Distributed Data Data sets,” Data Mining and Knowledge Discovery, vol. 12, no. 3, 2006.
[9] J. Von Neumann, “Various Techniques Used in Connection with Random Digits,” Nat'l Bureau of Standards Applied Math. Series 12, pp. 36-38, 1951.
[10] Y. Freund and R. Schapire, “Experiments with a New Boosting Algorithm,” Proc. 13th Int'l Conf. Machine Learning, 1996.
[11] A. Krieger, C. Long, and A. Wyner, “Boosting Noisy Data,” Proc. 18th Int'l Conf. Machine Learning, pp. 274-281, 2001.
[12] P. Chan and S. Stolfo, “Toward Scalable Learning with Non-Uniform Class and Cost Distributions,” Proc. Int'l Conf. Knowledge Discovery and Data Mining, 1998.
[13] J. Quinlan http://rulequest.comsee5-info.html, 1997.
[14] C. Blake and C. Merz UCI Data Repository, 1998.
[15] S. Hettich, and S. Bay The UCI KDD Archive, 1999.
[16] L. Breiman, “Bagging Predictors,” Machine Learning, vol. 24, no. 2, pp. 123-140, 1996.
[17] X. Zhu, and X. Wu, “Cost-Guided Class Noise Handling for Effective Cost-Sensitive Learning,” Proc. Fourth Int'l Conf. Data Mining, 2004.

Index Terms:
Data mining, classification, cost-sensitive learning, noise handling.
Xingquan Zhu, Xindong Wu, "Class Noise Handling for Effective Cost-Sensitive Learning by Cost-Guided Iterative Classification Filtering," IEEE Transactions on Knowledge and Data Engineering, vol. 18, no. 10, pp. 1435-1440, Oct. 2006, doi:10.1109/TKDE.2006.155
Usage of this product signifies your acceptance of the Terms of Use.