
This Article  
 
Share  
Bibliographic References  
Add to:  
Digg Furl Spurl Blink Simpy Del.icio.us Y!MyWeb  
Search  
 
ASCII Text  x  
Nizar Bouguila, "Clustering of Count Data Using Generalized Dirichlet Multinomial Distributions," IEEE Transactions on Knowledge and Data Engineering, vol. 20, no. 4, pp. 462474, April, 2008.  
BibTex  x  
@article{ 10.1109/TKDE.2007.190726, author = {Nizar Bouguila}, title = {Clustering of Count Data Using Generalized Dirichlet Multinomial Distributions}, journal ={IEEE Transactions on Knowledge and Data Engineering}, volume = {20}, number = {4}, issn = {10414347}, year = {2008}, pages = {462474}, doi = {http://doi.ieeecomputersociety.org/10.1109/TKDE.2007.190726}, publisher = {IEEE Computer Society}, address = {Los Alamitos, CA, USA}, }  
RefWorks Procite/RefMan/Endnote  x  
TY  JOUR JO  IEEE Transactions on Knowledge and Data Engineering TI  Clustering of Count Data Using Generalized Dirichlet Multinomial Distributions IS  4 SN  10414347 SP462 EP474 EPD  462474 A1  Nizar Bouguila, PY  2008 KW  clustering KW  Feature extraction KW  Image databases VL  20 JA  IEEE Transactions on Knowledge and Data Engineering ER   
[1] G.J. McLachlan and D. Peel, Finite Mixture Models. John Wiley & Sons, 2000.
[2] H. Tirri, P. Kontkanen, and P. Myllymäki, “Probabilistic InstanceBased Learning,” Proc. 13th Int'l Conf. Machine Learning (ICML '96), pp. 507515, 1996.
[3] K. Nigam, A. McCallum, S. Thrun, and T. Mitchell, “Text Classification from Labeled and Unlabeled Documents Using EM,” Machine Learning, vol. 39, no. 2, pp. 103134, 2000.
[4] M. Meilã and D. Heckerman, “An Experimental Comparison of ModelBased Clustering Methods,” Machine Learning, vol. 42, no. 12, pp. 929, 2001.
[5] F. Sebastiani, “Machine Learning in Automated Text Categorization,” ACM Computing Survey, vol. 34, pp. 147, 2002.
[6] P.S. Laplace, Philosophical Essay on Probabilities. Springer, 1995.
[7] D. Lewis, “Naive (Bayes) at Forty: The Independence Assumption in Information Retrieval,” Proc. 10th European Conf. Machine Learning (ECML '98), pp. 415, 1998.
[8] J.D.M. Rennie, L. Shih, J. Teevan, and D.R. Karger, “Tackling the Poor Assumptions of Naive Bayes Text Classifiers,” Proc. 20th Int'l Conf. Machine Learning (ICML '03), pp. 616623, 2003.
[9] R.E. Madsen, D. Kauchak, and C. Elkan, “Modeling Word Burstiness Using the Dirichlet Distribution,” Proc. 22nd Int'l Conf. Machine Learning (ICML '05), pp. 545552, 2005.
[10] J. Teevan and D.R. Karger, “Empirical Development of an Exponential Probabilistic Model for Text Retrieval: Using Textual Analysis to Build a Better Model,” Proc. ACM SIGIR '03, pp. 1825, 2003.
[11] M.H. DeGroot, Optimal Statistical Decisions. WileyInterscience, 2004.
[12] T.P. Minka, Estimating a Dirichlet Distribution, unpublished paper, http://research.microsoft.com/~minka/papers dirichlet/, 2007.
[13] D.J.C. Mackay and L. Peto, “A Hierarchical Dirichlet Language Model,” Natural Language Eng., vol. 1, no. 3, pp. 119, 1994.
[14] K. Sjolander, K. Karplus, M. Brown, R. Hughey, A. Krogh, I.S. Mian, and D. Haussler, “Dirichlet Mixtures: A Method for Improving Detection of Weak but Significant Protein Sequence Homology,” Computer Applications in the Biosciences, vol. 12, no. 4, pp. 327345, 1996.
[15] N. Bouguila and D. Ziou, “Unsupervised Learning of a Finite Discrete Mixture: Applications to Texture Modeling and Image Databases Summarization,” J. Visual Comm. and Image Representation, vol. 18, no. 4, pp. 295309, 2007.
[16] N. Bouguila and D. Ziou, “A Hybrid SEM Algorithm for HighDimensional Unsupervised Learning Using a Finite Generalized Dirichlet Mixture,” IEEE Trans. Image Processing, vol. 15, no. 9, pp.26572668, 2006.
[17] N. Bouguila and D. Ziou, “A Powerful Finite Mixture Model Based on the Generalized Dirichlet Distribution: Unsupervised Learning and Applications,” Proc. 17th Int'l Conf. Pattern Recognition (ICPR '04), pp. 280283, 2004.
[18] N. Bouguila and D. Ziou, “MMLBased Approach for HighDimensional Learning Using the Generalized Dirichlet Mixture,” Proc. IEEE CS Conf. Computer Vision and Pattern Recognition (CVPR '05) Workshops, vol. 3, p. 53, 2005.
[19] N. Bouguila and D. Ziou, “HighDimensional Unsupervised Selection and Estimation of a Finite Generalized Dirichlet Mixture Model Based on Minimum Message Length,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 29, no. 10, pp. 17161731, Oct. 2007.
[20] C. Elkan, “Clustering Documents with an ExponentialFamily Approximation of the Dirichlet Compound Multinomial Distribution,” Proc. 23rd Int'l Conf. Machine Learning (ICML '06), pp. 289296, 2006.
[21] K.W. Churchand and W.A. Gale, “Poisson Mixtures,” Natural Language Eng., vol. 1, pp. 163190, 1995.
[22] S.M. Katz, “Distribution of Content Words and Phrases in Text and Language Modelling,” Natural Language Eng., vol. 2, pp. 1559, 1996.
[23] N. Bouguila, D. Ziou, and J. Vaillancourt, “Unsupervised Learning of a Finite Mixture Model Based on the Dirichlet Distribution and Its Application,” IEEE Trans. Image Processing, vol. 13, no. 11, pp. 15331543, 2004.
[24] J.E. Mosimann, “On the Compound Multinomial Distribution, the Multivariate $\beta\hbox{}{\rm Distribution}$ , and Correlations among Proportions,” Biometrika, vol. 49, no. 12, pp. 6582, 1962.
[25] G. Pólya, “Sur Quelques Points de la Théorie des Probabilités,” Annales de l'Institut Henri Poincaré, vol. 1, no. 2, pp. 117161, 1930.
[26] J.G. Skellam, “A Probability Distribution Derived from the Binomial Distribution by Regarding the Probability of Success as Variable between the Sets of Trials,” J. Royal Statistical Soc., B, vol. 10, no. 2, pp. 257261, 1948.
[27] J.H. Darwin, “An Ecological Distribution Akin to Fisher's Logarithmic Distribution,” Biometrics, vol. 16, no. 1, pp. 5160, 1960.
[28] R.M. Dorazio and J.A. Royle, “Mixture Models for Estimating the Size of a Closed Population When Capture Rates Vary among Individuals,” Biometrics, vol. 59, no. 2, pp. 351364, 2003.
[29] R.H. Lochner, “A Generalized Dirichlet Distribution in Bayesian Life Testing,” J. Royal Statistical Soc., B, vol. 37, pp. 103113, 1975.
[30] R.J. Connor and J.E. Mosimann, “Concepts of Independence for Proportions with a Generalization of the Dirichlet Distribution,” J.Am. Statistical Assoc., vol. 64, pp. 194206, 1969.
[31] P.F. Thall and H.G. Sung, “Some Extensions and Applications of a Bayesian Strategy for Monitoring Multiple Outcomes in Clinical Trials,” Statistics in Medicine, vol. 17, pp. 15631580, 1998.
[32] P. Lewy, “A Generalized Dirichlet Distribution Accounting for Sigularities of the Variables,” Biometrics, vol. 52, no. 4, pp. 13941409, 1996.
[33] T. Wong, “Generalized Dirichlet Distribution in Bayesian Analysis,” Applied Math. and Computation, vol. 97, pp. 165181, 1998.
[34] N. Bouguila and D. Ziou, “Unsupervised Selection of a Finite Dirichlet Mixture Model: An MMLBased Approach,” IEEE Trans. Knowledge and Data Eng., vol. 18, no. 8, pp. 9931009, Aug. 2006.
[35] M.E. Tipping, “The Relevance Vector Machine,” Proc. Advances in Neural Information Processing Systems (NIPS '00), T.K. Leen, S.A.Solla, and K.R. Müller, eds., pp. 652658, 2000.
[36] I.J. Good, The Estimation of Probabilities. MIT Press, 1965.
[37] I.J. Good, Good Thinking: The Foundations of Probability and Its Applications. Univ. of Minneapolis Press, 1983.
[38] X. Meng and D. van Dyk, “The EM Algorithm—An Old Folk Song Sung to a Fast New Tune,” J. Royal Statistical Soc., B, vol. 59, no. 3, pp. 511567, 1997.
[39] A.P. Dempster, N.M. Laird, and D.B. Rubin, “Maximum Likelihood from Incomplete Data via the EM Algorithm,” J. Royal Statistical Soc., B, vol. 39, no. 1, pp. 138, 1977.
[40] G.J. McLachlan and T. Krishnan, The EM Algorithm and Extensions. WileyInterscience, 1997.
[41] N. Ueda and R. Nakano, “Deterministic Annealing EM Algorithm,” Neural Networks, vol. 11, pp. 271282, 1998.
[42] F.A. Graybill, Matrices with Applications in Statistics. Wadsworth, 1983.
[43] E. Hille, Analytic Function Theory, vol. 1. Ginn and Co., 1959.
[44] K. Lange, “Applications of the Dirichlet Distribution to Forensic Match Probabilities,” Genetica, vol. 96, pp. 107117, 1995.
[45] J. Rissanen, “Modeling by Shortest Data Description,” Automatica, vol. 14, pp. 465471, 1987.
[46] Y. Rui, T.S. Huang, and S. Chang, “Image Retrieval: Current Techniques, Promising Directions, and Open Issues,” J. Visual Comm. and Image Representation, vol. 10, no. 4, pp. 3962, Apr. 1999.
[47] A.W.M. Smeulders, M. Worring, S. Santini, A. Gupta, and R. Jain, “ContentBased Image Retrieval at the End of the Early Years,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 22, no. 12, pp. 13491380, Dec. 2000.
[48] W.W. Chu, C.C. Hsu, A.F. Cardenasand, and R.K. Taira, “KnowledgeBased Image Retrieval with Spatial and Temporal Constructs,” IEEE Trans. Knowledge and Data Eng., vol. 10, no. 6, pp.872888, Nov./Dec. 1998.
[49] G. Lu, “Techniques and Data Structures for Efficient Multimedia Retrieval Based on Similarity,” IEEE Trans. Multimedia, vol. 4, no. 3, pp. 372384, 2002.
[50] M.S. Lew, “NextGeneration Web Searches for Visual Content,” Computer, vol. 33, pp. 4653, 2000.
[51] M. Swain and D. Ballard, “Color Indexing,” Int'l J. Computer Vision, vol. 7, no. 1, pp. 1132, 1991.
[52] A. Pentland, R. Picard, and S. Sclaroff, “Photobook: ContentBased Manipulation of Image Databases,” Int'l J. Computer Vision, vol. 18, no. 3, pp. 233254, 1996.
[53] H.Y. Lee, H.K. Lee, and H.H. Yeong, “Spatial Color Descriptor for Image Retrieval and Video Segmentation,” IEEE Trans. Multimedia, vol. 5, no. 3, pp. 358367, 2003.
[54] G. Pass and R. Zabih, “Comparing Images Using Joint Histograms,” ACM J. Multimedia Systems, vol. 7, no. 3, pp. 234240, 1999.
[55] Y. Tao and W.I. Grosky, “Spatial Color Indexing Using Rotation, Translation, and Scale Invariant Anglograms,” Multimedia Tools and Applications, vol. 15, no. 3, pp. 247268, 2001.
[56] J. Huang, S.R. Kumar, M. Mitra, W. Zhu, and R. Zabih, “Spatial Color Indexing and Applications,” Int'l J. Computer Vision, vol. 35, no. 3, pp. 245268, 1999.
[57] R.M. Haralick, K. Shanmugan, and I. Dinstein, “Textural Features for Image Classification,” IEEE Trans. Systems, Man and Cybernetics, vol. 8, pp. 610621, 1973.
[58] G. Celeux and G. Govaert, “A Classification EM Algorithm for Clustering and Two Stochastic Versions,” Computational Statistics and Data Analysis, vol. 14, pp. 315332, 1992.
[59] G. Celeux and G. Govaert, “Comparison of the Mixture and the Classification Maximum Likelihood in Cluster Analysis,” J.Statistical Computation and Simulation, vol. 47, pp. 127146, 1993.
[60] G. Celeux and G. Govaert, “Gaussian Parsimonious Clustering Models,” Pattern Recognition, vol. 28, no. 5, pp. 781793, 1995.
[61] C. Biernacki, G. Celeux, and G. Govaert, “Assessing a Mixture Model for Clustering with the Integrated Complete Likelihood,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 22, no. 7, pp. 719725, July 2000.
[62] C. Biernacki, G. Celeux, and G. Govaert, “An Improvement of the NEC Criterion for Assessing the Number of Clusters in a Mixture Model,” Pattern Recognition Letters, vol. 20, pp. 267272, 1999.
[63] A. Agresti, Categorical Data Analysis. John Wiley & Sons, 2002.
[64] R. Plamondon and S.N. Srihari, “OnLine and OffLine Handwriting Recognition: A Comprehensive Survey,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 22, no. 1, pp. 6384, Jan. 2000.
[65] P.M. Murphy and D.W. Aha,, UCI Repository of Machine Learning Databases, Dept. of Information and Computer Science, Univ. of California, Irvine, http://www.ics.ci.edu/mlearnMLRepository. html. , 1998.
[66] D. Verma and M. Meilã, “A Comparison of Spectral Clustering Algorithms,” Technical Report uwcse030501, Univ. Washington, 2003.
[67] A.L.I. Oliveira, C.A.B. Mello, E.R. Silva Jr, and V.M.O. Alves, “Optical Digit Recognition for Images of Handwritten Documents,” Proc. Ninth Brazilian Symp. Neural Networks (SBRN '06), p.92, 2006.
[68] D. Keysers, C. Gollan, and H. Ney, “Local Context in NonLinear Deformation Models for Handwritten Character Recognition,” Proc. 17th Int'l Conf. Pattern Recognition (ICPR '04), pp. 511514, 2004.
[69] A.K. McCallum, “Bow: A Toolkit for Statistical Language Modeling, Text Retrieval, Classification and Clustering,” http://www.cs.cmu.edu/mccallumbow, 2007.