The Community for Technology Leaders
RSS Icon
Subscribe
Issue No.01 - January (2010 vol.22)
pp: 76-89
Wu-Jun Li , Hong Kong University of Science and Technology, Hong Kong
Dit-Yan Yeung , Hong Kong University of Science and Technology, Hong Kong
ABSTRACT
In multiple-instance learning (MIL), an individual example is called an instance and a bag contains a single or multiple instances. The class labels available in the training set are associated with bags rather than instances. A bag is labeled positive if at least one of its instances is positive; otherwise, the bag is labeled negative. Since a positive bag may contain some negative instances in addition to one or more positive instances, the true labels for the instances in a positive bag may or may not be the same as the corresponding bag label and, consequently, the instance labels are inherently ambiguous. In this paper, we propose a very efficient and robust MIL method, called Multiple-Instance Learning via Disambiguation (MILD), for general MIL problems. First, we propose a novel disambiguation method to identify the true positive instances in the positive bags. Second, we propose two feature representation schemes, one for instance-level classification and the other for bag-level classification, to convert the MIL problem into a standard single-instance learning (SIL) problem that can be solved by well-known SIL algorithms, such as support vector machine. Third, an inductive semi-supervised learning method is proposed for MIL. We evaluate our methods extensively on several challenging MIL applications to demonstrate their promising efficiency, robustness, and accuracy.
INDEX TERMS
Multiple-instance learning, learning from ambiguity, CBIR, object recognition, co-training, drug activity prediction.
CITATION
Wu-Jun Li, Dit-Yan Yeung, "MILD: Multiple-Instance Learning via Disambiguation", IEEE Transactions on Knowledge & Data Engineering, vol.22, no. 1, pp. 76-89, January 2010, doi:10.1109/TKDE.2009.58
REFERENCES
[1] T.G. Dietterich, R.H. Lathrop, and T. Lozano-Pérez, “Solving the Multiple Instance Problem with Axis-Parallel Rectangles,” Artificial Intelligence, vol. 89, nos. 1/2, pp. 31-71, 1997.
[2] S. Ray and M. Craven, “Supervised versus Multiple Instance Learning: An Empirical Comparison,” Proc. Int'l Conf. Machine Learning, 2005.
[3] X. Xu, “Statistical Learning in Multiple Instance Problem,” master's thesis, Univ. of Waikato, 0657.594, 2003.
[4] Y. Chen and J.Z. Wang, “Image Categorization by Learning and Reasoning with Regions,” J. Machine Learning Research, vol. 5, pp.913-939, 2004.
[5] R. Rahmani and S.A. Goldman, “MISSL: Multiple-Instance Semi-Supervised Learning,” Proc. Int'l Conf. Machine Learning, 2006.
[6] T. Gärtner, P.A. Flach, A. Kowalczyk, and A.J. Smola, “Multi-Instance Kernels,” Proc. Int'l Conf. Machine Learning, 2002.
[7] S. Andrews, I. Tsochantaridis, and T. Hofmann, “Support Vector Machines for Multiple-Instance Learning,” Proc. Conf. Advances in Neural Information Processing Systems, pp. 561-568, 2002.
[8] P.-M. Cheung and J.T. Kwok, “A Regularization Framework for Multiple-Instance Learning,” Proc. Int'l Conf. Machine Learning, 2006.
[9] Z.-H. Zhou and J.-M. Xu, “On the Relation between Multi-Instance Learning and Semi-Supervised Learning,” Proc. Int'l Conf. Machine Learning, 2007.
[10] O. Maron and T. Lozano-Pérez, “A Framework for Multiple-Instance Learning,” Proc. Conf. Advances in Neural Information Processing Systems, 1997.
[11] Y. Chen, J. Bi, and J.Z. Wang, “MILES: Multiple-Instance Learning via Embedded Instance Selection,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 28, no. 12, pp. 1931-1947, Dec. 2006.
[12] A. Blum and T.M. Mitchell, “Combining Labeled and Unlabeled Data with Co-Training,” Proc. Ann. Conf. Learning Theory, pp. 92-100, 1998.
[13] J. Davis, V.S. Costa, S. Ray, and D. Page, “An Integrated Approach to Feature Invention and Model Construction for Drug Activity Prediction,” Proc. Int'l Conf. Machine Learning, 2007.
[14] P. Auer, “On Learning from Multi-Instance Examples: Empirical Evaluation of a Theoretical Approach,” Proc. Int'l Conf. Machine Learning, pp. 21-29, 1997.
[15] J. Ramon and L.D. Raedt, “Multi-Instance Neural Networks,” Proc. Int'l Conf. Machine Learning (ICML '00) Workshop Attribute-Value and Relational Learning, 2000.
[16] Z.-H. Zhou and M.-L. Zhang, “Multi-Instance Multi-Label Learning with Application to Scene Classification,” Proc. Conf. Advances in Neural Information Processing Systems, pp. 1609-1616, 2006.
[17] J. Bi, Y. Chen, and J.Z. Wang, “A Sparse Support Vector Machine Approach to Region-Based Image Categorization,” Proc. IEEE CS Conf. Computer Vision and Pattern Recognition, vol. 1, pp. 1121-1128, 2005.
[18] O. Maron and A.L. Ratan, “Multiple-Instance Learning for Natural Scene Classification,” Proc. Int'l Conf. Machine Learning, pp. 341-349, 1998.
[19] Q. Tao, S. Scott, N.V. Vinodchandran, and T.T. Osugi, “SVM-Based Generalized Multiple-Instance Learning via Approximate Box Counting,” Proc. Int'l Conf. Machine Learning, 2004.
[20] J. Bi and J. Liang, “Multiple Instance Learning of Pulmonary Embolism Detection with Geodesic Distance Along Vascular Structure,” Proc. IEEE CS Conf. Computer Vision and Pattern Recognition, 2007.
[21] G. Fung, M. Dundar, B. Krishnapuram, and R.B. Rao, “Multiple Instance Learning for Computer Aided Diagnosis,” Proc. Conf. Advances in Neural Information Processing Systems, 2006.
[22] G. Ruffo, “Learning Single and Multiple Decision Trees for Security Applications,” PhD dissertation, Univ. of Turin, 2000.
[23] Q. Zhang and S.A. Goldman, “EM-DD: An Improved Multiple-Instance Learning Technique,” Proc. Conf. Advances in Neural Information Processing Systems, 2001.
[24] Q. Zhang, S.A. Goldman, W. Yu, and J. Fritts, “Content-Based Image Retrieval Using Multiple-Instance Learning,” Proc. Int'l Conf. Machine Learning, pp. 682-689, 2002.
[25] A. Dempster, N. Laird, and D. Rubin, “Maximum Likelihood from Incomplete Data via the EM Algorithm,” J. Royal Statistical Soc., vol. 39, no. 1, pp. 1-38, 1977.
[26] P.V. Gehler and O. Chapelle, “Deterministic Annealing for Multiple-Instance Learning,” Proc. Int'l Conf. Artificial Intelligence and Statistics, 2007.
[27] J.T. Kwok and P.-M. Cheung, “Marginalized Multi-Instance Kernels,” Proc. Int'l Joint Conf. Artificial Intelligence, 2007.
[28] R.C. Bunescu and R.J. Mooney, “Multiple Instance Learning for Sparse Positive Bags,” Proc. Int'l Conf. Machine Learning, 2007.
[29] R.O. Duda, P.E. Hart, and D.G. Stork, Pattern Classification, second ed. Wiley Interscience, 2000.
[30] N. Cristianini and J. Shawe-Taylor, An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods. Cambridge Univ. Press, 2000.
[31] S. Rüping and T. Scheffer, “Learning with Multiple Views,” Proc. Int'l Conf. Machine Learning (ICML) Workshop Learning with Multiple Views, 2005.
[32] A. Opelt, A. Pinz, M. Fussenegger, and P. Auer, “Generic Object Recognition with Boosting,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 28, no. 3, pp. 416-431, Mar. 2006.
[33] R. Rahmani, S.A. Goldman, H. Zhang, J. Krettek, and J.E. Fritts, “Localized Content Based Image Retrieval,” Proc. Multimedia Information Retrieval, pp. 227-236, 2005.
[34] F. Perronnin, “Universal and Adapted Vocabularies for Generic Visual Categorization,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 30, no. 7, pp. 1243-1256, July 2008.
[35] C.-C. Chang and C.-J. Lin, “LIBSVM: A Library for Support Vector Machines,” http://www.csie.ntu.edu.tw/cjlinlibsvm, 2001.
[36] C. Wang, S.D. Scott, J. Zhang, Q. Tao, D.E. Fomenko, and V.N. Gladyshev, “A Study in Modeling Low-Conservation Protein Superfamilies,” Technical Report UNL-CSE-2004-0003, Univ. of Nebraska, 2004.
[37] B. Settles, M. Craven, and S. Ray, “Multiple-Instance Active Learning,” Proc. Conf. Advances in Neural Information Processing Systems, 2007.
[38] J. Sivic and A. Zisserman, “Video Google: A Text Retrieval Approach to Object Matching in Videos,” Proc. Int'l Conf. Computer Vision (ICCV), pp. 1470-1477, 2003.
[39] J.R. Quinlan, “Learning Logical Definitions from Relations,” Machine Learning, vol. 5, pp. 239-266, 1990.
[40] A. Argyriou, M. Herbster, and M. Pontil, “Combining Graph Laplacians for Semi-Supervised Learning,” Proc. Conf. Advances in Neural Information Processing Systems, 2005.
5 ms
(Ver 2.0)

Marketing Automation Platform Marketing Automation Tool