The Community for Technology Leaders
RSS Icon
Subscribe
Issue No.07 - July (2009 vol.31)
pp: 1195-1209
Ying Wu , Northwestern University, Evanston
Gang Hua , Microsoft Research, Redmond
ABSTRACT
Enormous uncertainties in unconstrained environments lead to a fundamental dilemma that many tracking algorithms have to face in practice: Tracking has to be computationally efficient, but verifying whether or not the tracker is following the true target tends to be demanding, especially when the background is cluttered and/or when occlusion occurs. Due to the lack of a good solution to this problem, many existing methods tend to be either effective but computationally intensive by using sophisticated image observation models or efficient but vulnerable to false alarms. This greatly challenges long-duration robust tracking. This paper presents a novel solution to this dilemma by considering the context of the tracking scene. Specifically, we integrate into the tracking process a set of auxiliary objects that are automatically discovered in the video on the fly by data mining. Auxiliary objects have three properties, at least in a short time interval: 1) persistent co-occurrence with the target, 2) consistent motion correlation to the target, and 3) easy to track. Regarding these auxiliary objects as the context of the target, the collaborative tracking of these auxiliary objects leads to efficient computation as well as strong verification. Our extensive experiments have exhibited exciting performance in very challenging real-world testing cases.
INDEX TERMS
Computer vision, visual object tracking, context aware, collaborative tracking, data mining, robust fusion, belief inconsistency.
CITATION
Ying Wu, Gang Hua, "Context-Aware Visual Tracking", IEEE Transactions on Pattern Analysis & Machine Intelligence, vol.31, no. 7, pp. 1195-1209, July 2009, doi:10.1109/TPAMI.2008.146
REFERENCES
[1] M. Isard and A. Blake, “Contour Tracking by Stochastic Propagation of Conditional Density,” Proc. Fourth European Conf. Computer Vision, pp. 343-356, Apr. 1996.
[2] S. Birchfield, “Elliptical Head Tracking Using Intensity Gradients and Color Histograms,” Proc. IEEE Conf. Computer Vision and Pattern Recognition, pp. 232- 237, June 1998.
[3] D. Comaniciu, V. Ramesh, and P. Meer, “Kernel-Based Object Tracking,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 25, no. 5, pp. 564-577, May 2003.
[4] M.J. Black and A.D. Jepson, “Eigentracking: Robust Matching and Tracking of Articulated Objects Using a View-Based Representation,” Proc. Fourth European Conf. Computer Vision, pp. 329-342, Apr. 1996.
[5] G. Hager and P. Belhumeur, “Real-Time Tracking of Image Regions with Changes in Geometry and Illumination,” Proc. IEEE Conf. Computer Vision and Pattern Recognition, pp. 403-410, June 1996.
[6] G. Hager, M. Dewan, and C. Stewart, “Multiple Kernel Tracking with SSD,” Proc. IEEE Conf. Computer Vision and Pattern Recognition, vol. 1, pp. 790-797, June/July 2004.
[7] K.-C. Lee and D. Kriegman, “Online Learning of Probabilistic Appearance Manifolds for Video-Based Recognition and Tracking,” Proc. IEEE Conf. Computer Vision and Pattern Recognition, vol. 1, pp. 852-859, June 2005.
[8] J. Lim, D. Ross, R.-S. Lin, and M.-H. Yang, “Incremental Learning for Visual Tracking,” Proc. Advances in Neural Information Processing Systems, vol. 17, pp. 801-808, Dec. 2004.
[9] M. Yang and Y. Wu, “Tracking Non-Stationary Appearances and Dynamic Feature Selection,” Proc. IEEE Conf. Computer Vision and Pattern Recognition, vol. 2, pp. 1059-1066, June 2005.
[10] S. Avidan, “Support Vector Tracking,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 26, no. 8, pp. 1064-1072, Aug. 2004.
[11] S. Avidan, “Ensemble Tracking,” Proc. IEEE Conf. Computer Vision and Pattern Recognition, vol. 2, pp. 494-501, June 2005.
[12] J. Wang, X. Chen, and W. Gao, “Online Selecting Discriminative Tracking Features Using Particle Filter,” Proc. IEEE Conf. Computer Vision and Pattern Recognition, vol. 2, pp. 1037-1042, June 2005.
[13] R. Agrawal and R. Srikant, “Fast Algorithms for Mining Association Rules,” Proc. 20th Int'l Conf. Very Large Data Bases, pp. 487-499, 1994.
[14] J. Sivic and A. Zisserman, “Video Google: A Text Retrieval Approach to Object Matching in Videos,” Proc. Ninth IEEE Int'l Conf. Computer Vision, vol. 2, pp. 1470-1477, Oct. 2003.
[15] J. Sivic and A. Zisserman, “Video Data Mining Using Configurations of Viewpoint Invariant Regions,” Proc. IEEE Conf. Computer Vision and Pattern Recognition, vol. 1, pp. 488-495, June/July 2004.
[16] M. Leordeanu and R. Collins, “Unsupervised Learning of Object Features from Video Sequences,” Proc. IEEE Conf. Computer Vision and Pattern Recognition, vol. 1, pp. 1142-1149, June 2005.
[17] X.S. Zhou, D. Comaniciu, and A. Gupta, “An Information Fusion Framework for Robust Shape Tracking,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 27, no. 1, pp. 115-129, Jan. 2005.
[18] D.G.C. Race, http://www.darpa.milgrandchallenge, 2008.
[19] B.K. Horn and B.G. Schunck, “Determining Optical Flow,” Artificial Intelligence, vol. 17, pp. 185-203, 1981.
[20] B.D. Lucas and T. Kanade, “An Iterative Image Registration Technique with an Application to Stereo Vision,” Proc. DARPA Image Understanding Workshop, pp. 121-130, Apr. 1981.
[21] T. Brox, A. Bruhn, N. Papenberg, and J. Weickert, “High Accuracy Optical Flow Estimation Based on a Theory for Warping,” Proc. Eighth European Conf. Computer Vision, vol. 4, pp. 25-36, May 2004.
[22] S. Roth and M.J. Black, “On the Spatial Statistics of Optical Flow,” Proc. 10th IEEE Int'l Conf. Computer Vision, vol. 1, pp. 42-49, Oct. 2005.
[23] A. Jepson, D. Fleet, and T. El-Maraghi, “Robust Online Appearance Models for Visual Tracking,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 25, no. 10, pp. 1296-1311, Oct. 2003.
[24] D. Ramanan, D.A. Forsyth, and A. Zisserman, “Tracking People by Learning Their Appearance,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 29, no. 1, pp. 65-81, Jan. 2007.
[25] K. Okuma, A. Taleghani, N.D. Freitas, J.J. Little, and D.G. Lowe, “A Boosted Particle Filter: Multitarget Detection and Tracking,” Proc. Eighth European Conf. Computer Vision, vol. 1, pp. 28-39, May 2004.
[26] T. Zhao and R. Nevatia, “Tracking Multiple Humans in Complex Situations,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 26, no. 9, pp. 1208-1221, Sept. 2004.
[27] C. Stauffer and W. Grimson, “Adaptive Background Mixture Models for Real-Time Tracking,” Proc. IEEE Conf. Computer Vision and Pattern Recognition, vol. 2, pp. 246-252, June 1999.
[28] J. Shi and J. Malik, “Motion Segmentation and Tracking Using Normalized Cuts,” Proc. Sixth IEEE Int'l Conf. Computer Vision, pp.1154-1160, Jan. 1998.
[29] O. Williams, A. Blake, and R. Cipolla, “Sparse Bayesian Learning for Efficient Visual Tracking,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 27, no. 8, pp. 1292-1304, Aug. 2005.
[30] A. Yilmaz, O. Javed, and M. Shah, “Object Tracking: A Survey,” ACM Computing Surveys, vol. 38, no. 4, p. 13, Dec. 2006.
[31] M. Isard and A. Blake, “CONDENSATION—Conditional Density Propagation for Visual Tracking,” Int'l J. Computer Vision, vol. 29, pp. 5-28, May 1998.
[32] D.B. Reid, “An Algorithm for Tracking Multiple Targets,” IEEE Trans. Automatic Control, vol. 24, no. 6, pp. 843-854, Dec. 1979.
[33] I.J. Cox and S.L. Hingorani, “An Efficient Implementation of Reid's Multiple Hypothesis Tracking Algorithm and Its Evaluation for the Purpose of Visual Tracking,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 18, no. 2, pp. 138-150, Feb. 1996.
[34] Y. Bar-Shalom and T.E. Fortmann, Tracking and Data Association. Academic Press, 1988.
[35] C. Rasmussen and G.D. Hager, “Probabilistic Data Association Methods for Tracking Complex Visual Objects,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 23, no. 6, pp. 560-576, June 2001.
[36] Y. Wu and T.S. Huang, “Robust Visual Tracking by Integrating Multiple Cues Based on Co-Inference Learning,” Int'l J. Computer Vision, vol. 58, no. 1, pp. 55-71, June 2004.
[37] R.T. Collins, Y. Liu, and M. Leordeanu, “Robust Online Appearance Models for Visual Tracking,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 27, no. 10, pp. 1631-1643, Oct. 2005.
[38] H. Grabner, M. Grabner, and H. Bischof, “Real-Time Tracking via On-Line Boosting,” Proc. British Machine Vision Conf., vol. 1, 4-7, pp. 47-56, 2006.
[39] M. Yang, Y. Wu, and S. Lao, “Intelligent Collaborative Tracking by Mining Auxiliary Objects,” Proc. IEEE Conf. Computer Vision and Pattern Recognition, vol. 1, pp. 697-704, June 2006.
[40] J. Shi and C. Tomasi, “Good Features to Track,” Proc. IEEE Conf. Computer Vision and Pattern Recognition, pp. 593-600, June 1994.
[41] D.G. Lowe, “Object Recognition from Local Scale-Invariant Features,” Proc. Seventh IEEE Int'l Conf. Computer Vision, vol. 2, pp. 1150-1157, Sept. 1999.
[42] K. Mikolajczyk and C. Schmid, “Indexing Based on Scale Invariant Interest Points,” Proc. Eighth IEEE Int'l Conf. Computer Vision, vol. 1, pp. 525-531, July 2001.
[43] A. Fitzgibbon and A. Zisserman, “On Affine Invariant Clustering and Automatic Cast Listing in Movies,” Proc. Seventh European Conf. Computer Vision, vol. 3, pp. 304-320, May/June 2002.
[44] R. Jain, R. Kasturi, and B.G. Schunck, Machine Vision. McGraw-Hill, 1995.
[45] S.J. Julier and J.K. Uhlmann, “A Non-Divergent Estimation Algorithm in the Presence of Unknown Correlations,” Proc. Am. Control Conf., pp. 2369-2373, June 1997.
[46] G. Hua and Y. Wu, “Measurement Integration under Inconsistency for Robust Tracking,” Proc. IEEE Conf. Computer Vision and Pattern Recognition, vol. 1, pp. 650-657, June 2006.
[47] EC Funded CAVIAR Project/IST 2001 37540, http://homepages.inf.ed.ac.uk/rbfcaviar/, 2008.
[48] V.I. Pavlovic, “Dynamic Bayesian Networks for Information Fusion with Application to Human-Computer Interfaces,” PhD dissertation, Dept. of Electrical and Computer Eng., Univ. of Illinois at Urbana-Champaign, 1999.
29 ms
(Ver 2.0)

Marketing Automation Platform Marketing Automation Tool