
This Article  
 
Share  
Bibliographic References  
Add to:  
Digg Furl Spurl Blink Simpy Del.icio.us Y!MyWeb  
Search  
 
ASCII Text  x  
Mark Last, Oded Maimon, "A Compact and Accurate Model for Classification," IEEE Transactions on Knowledge and Data Engineering, vol. 16, no. 2, pp. 203215, February, 2004.  
BibTex  x  
@article{ 10.1109/TKDE.2004.1269598, author = {Mark Last and Oded Maimon}, title = {A Compact and Accurate Model for Classification}, journal ={IEEE Transactions on Knowledge and Data Engineering}, volume = {16}, number = {2}, issn = {10414347}, year = {2004}, pages = {203215}, doi = {http://doi.ieeecomputersociety.org/10.1109/TKDE.2004.1269598}, publisher = {IEEE Computer Society}, address = {Los Alamitos, CA, USA}, }  
RefWorks Procite/RefMan/Endnote  x  
TY  JOUR JO  IEEE Transactions on Knowledge and Data Engineering TI  A Compact and Accurate Model for Classification IS  2 SN  10414347 SP203 EP215 EPD  203215 A1  Mark Last, A1  Oded Maimon, PY  2004 KW  Knowledge discovery in databases KW  data mining KW  classification KW  dimensionality reduction KW  feature selection KW  decision trees KW  information theory KW  Information theoretic network. VL  16 JA  IEEE Transactions on Knowledge and Data Engineering ER   
Abstract—We describe and evaluate an informationtheoretic algorithm for datadriven induction of classification models based on a minimal subset of available features. The relationship between input (predictive) features and the target (classification) attribute is modeled by a treelike structure termed an
[1] F. Attneave, Applications of Information Theory to Psychology. Holt, Rinehart, and Winston, 1959.
[2] C.L. Blake and C.J. Merz , UCI Repository of Machine Learning Databases,http://www.ics.uci.edu/~mlearnMLRepository. html , 19 July 2002.
[3] L. Breiman, J.H. Friedman, R.A. Olshen, and P.J. Stone, Classification and Regression Trees. Wadsworth, 1984.
[4] T.M. Cover and J.A. Thomas, Elements of Information Theory. Wiley, 1991.
[5] P. Domingos and M. Pazzani, On the Optimality of the Simple Bayesian Classifier under ZeroOne Loss Machine Learning, no. 29, pp. 103130, 1997.
[6] P. Domingos, Occam's Two Razors: The Sharp and the Blunt Proc. Fourth Int'l Conf. Knowledge Discovery and Data Mining, pp. 3743, 1998.
[7] U. Fayyad and K. Irani, MultiInterval Discretization of ContinuousValued Attributes for Classification Learning Proc. 13th Int'l Joint Conf. Artificial Intelligence, pp. 10221027, 1993.
[8] U. Fayyad, G. PiatetskyShapiro, and P. Smyth, From Data Mining to Knowledge Discovery: An Overview Advances in Knowledge Discovery and Data Mining, U. Fayyad, G. PiatetskyShapiro, P. Smyth, and R. Uthurusamy, eds., pp. 136, AAAI/MIT Press, 1996.
[9] A.L. Gorin, S.E. Levinson, A.N. Gertner, and E. Goldman, Adaptive Acquisition of Language Computer Speech and Language, vol. 5, no. 2, pp. 101132, 1991.
[10] A.L. Gorin, S.E. Levinson, and A. Sankar, An Experiment in Spoken Language Acquisition IEEE Trans. Speech and Audio Processing, vol. 2, no. 1, pp. 224239, 1994.
[11] G.H. John, R. Kohavi, and K. Pfleger, Irrelevant Features and the Subset Selection Problem Proc. 11th Int'l Conf. Machine Learning, pp. 121129, 1994.
[12] M. Last, Y. Klein, and A. Kandel, Knowledge Discovery in Time Series Databases IEEE Trans. Systems, Man, and Cybernetics, Part B, vol. 31, no. 1, pp. 160169, Feb. 2001.
[13] M. Last, A. Kandel, and O. Maimon, InformationTheoretic Algorithm for Feature Selection Pattern Recognition Letters, vol. 22, nos. 67, pp. 799811, 2001.
[14] M. Last and A. Kandel, Data Mining for Process and Quality Control in the Semiconductor Industry Data Mining for Design and Manufacturing: Methods and Applications, D. Braha, ed., pp. 207234, Boston: Kluwer Academic, 2001.
[15] H. Liu and H. Motoda, Feature Selection for Knowledge Discovery and Data Mining. Boston: Kluwer Academic, 1998.
[16] O. Maimon and M. Last, Knowledge Discovery and Data Mining, The InfoFuzzy Network (IFN) Methodology. Boston: Kluwer Academic, 2001.
[17] E.W. Minium, R.B. Clarke, and T. Coladarci, Elements of Statistical Reasoning. New York: Wiley, 1999.
[18] T.M. Mitchell, Machine Learning. McGrawHill, 1997.
[19] J.R. Quinlan, Induction of Decision Trees Machine Learning, vol. 1, no. 1, pp. 81106, 1986.
[20] J.R. Quinlan, Simplifying Decision Trees Int'l J. ManMachine Studies, no. 27, pp. 221234, 1987.
[21] J.R. Quinlan, C4.5: Programs for Machine Learning. Morgan Kaufmann, 1993.
[22] J.R. Quinlan, Improved Use of Continuous Attributes in C4.5 J. Artificial Intelligence Research, no. 4, pp. 7790, 1996.
[23] C.R. Rao and H. Toutenburg, Linear Models: Least Squares and Alternatives. SpringerVerlag, 1995.
[24] R. Rastogi and K. Shim, PUBLIC: A Decision Tree Classifier that Integrates Building and Pruning Proc. 24th Int'l Conf.Very Large Databases (VLDB '98), pp. 404415, 1998.
[25] P. Smyth and R. Goodman, "An Information Theoretic Approach to Rule Induction from Databases," IEEE Trans Knowledge and Data Eng., vol. 4, no. 4, pp. 301316, Aug. 1992.