The Community for Technology Leaders
RSS Icon
Subscribe
Issue No.05 - May (2013 vol.35)
pp: 1263-1273
R. Sznitman , EPFL IC ISIM CVLAB, Lausanne, Switzerland
R. Richa , Johns Hopkins Univ., Baltimore, MD, USA
R. H. Taylor , Johns Hopkins Univ., Baltimore, MD, USA
B. Jedynak , Johns Hopkins Univ., Baltimore, MD, USA
G. D. Hager , Johns Hopkins Univ., Baltimore, MD, USA
ABSTRACT
Methods for tracking an object have generally fallen into two groups: tracking by detection and tracking through local optimization. The advantage of detection-based tracking is its ability to deal with target appearance and disappearance, but it does not naturally take advantage of target motion continuity during detection. The advantage of local optimization is efficiency and accuracy, but it requires additional algorithms to initialize tracking when the target is lost. To bridge these two approaches, we propose a framework for unified detection and tracking as a time-series Bayesian estimation problem. The basis of our approach is to treat both detection and tracking as a sequential entropy minimization problem, where the goal is to determine the parameters describing a target in each frame. To do this we integrate the Active Testing (AT) paradigm with Bayesian filtering, and this results in a framework capable of both detecting and tracking robustly in situations where the target object enters and leaves the field of view regularly. We demonstrate our approach on a retinal tool tracking problem and show through extensive experiments that our method provides an efficient and robust tracking solution.
INDEX TERMS
Instruments, Target tracking, Surgery, Testing, Optimization, Aerospace electronics,retinal microsurgery, Unified object detection and tracking, active testing, instrument tracking, adaptive sensing
CITATION
R. Sznitman, R. Richa, R. H. Taylor, B. Jedynak, G. D. Hager, "Unified Detection and Tracking of Instruments during Retinal Microsurgery", IEEE Transactions on Pattern Analysis & Machine Intelligence, vol.35, no. 5, pp. 1263-1273, May 2013, doi:10.1109/TPAMI.2012.209
REFERENCES
[1] A. Yilmaz, O. Javed, and M. Shah, "Object Tracking: A Survey," ACM Computing Surveys, vol. 38, no. 4, 2006.
[2] D. Comaniciu, V. Ramesh, and P. Meer, "Real-Time Tracking of Non-Rigid Objects Using Mean Shift," Proc. IEEE Conf. Computer Vision and Pattern Recognition, pp. 142-149, 2000.
[3] D. Geman and B. Jedynak, "An Active Testing Model for Tracking Roads from Satellite Images," IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 18, no. 1, pp. 1-14, Jan. 1996.
[4] O. Bernier, P. Cheung Mon Chan, and A. Bouguet, "Fast Nonparametric Belief Propagation for Real-Time Stereo Articulated Body Tracking," J. Computer Vision and Image Understanding, vol. 113, no. 1, pp. 29-47, 2009.
[5] W. Geng, P. Cosman, C.C. Berry, Z. Feng, and W.R. Schafer, "Automatic Tracking, Feature Extraction and Classification of C Elegans Phenotypes," IEEE Trans. Biomedical Eng., vol. 51, no. 10, pp. 1811-1820, Oct. 2004.
[6] P. Viola and M. Jones, "Robust Real-Time Face Detection," Int'l J. Computer Vision, vol. 57, no. 2, pp. 137-154, 2004.
[7] A. Vedaldi, G. Gulshan, M. Varma, and A. Zisserman, "Multiple Kernels for Object Detection," Proc. IEEE Int'l Conf. Computer Vision, pp. 606-613, 2009.
[8] C. Lampert, M. Blaschko, and T. Hofmann, "Beyond Sliding Windows: Object Localization by Efficient Subwindow Search," Proc. IEEE Conf. Computer Vision and Pattern Recognition, pp. 1-8, 2008.
[9] D. Aldavert, A. Ramisa, R. Toledo, and R. Mantaras, "Fast and Robust Object Segmentation with the Integral Linear Classifier," Proc. IEEE Conf. Computer Vision and Pattern Recognition, pp. 1046-1053, 2010.
[10] R. Kalman, "A New Approach to Linear Filtering and Prediction Problems," J. Basic Eng., vol. 82, pp. 35-45, 1960.
[11] M. Isard and A. Blake, "Condensation-Conditional Density Propagation for Visual Tracking," Int'l J. Computer Vision, vol. 28, no. 1, pp. 5-28, 1998.
[12] M. Andriluka, S. Roth, and B. Schiele, "People-Tracking-by-Detection and People-Detection-by-Tracking," Proc. IEEE Conf. Computer Vision and Pattern Recognition, pp. 1-8, 2008.
[13] M. Breitenstein, F. Reichlin, B. Leibe, E. Koller-Meier, and L. Van Gool, "Robust Tracking-by-Detection Using a Detector Confidence Particle Filter," Proc. IEEE Int'l Conf. Computer Vision, pp. 1515-1522, 2009.
[14] S.S. Nejhum, J. Ho, and Y. Ming-Hsuan, "Visual Tracking with Histograms and Articulating Blocks," Proc. IEEE Conf. Computer Vision and Pattern Recognition, pp. 1-8, 2008.
[15] R. Verma, C. Schmid, and K. Mikolajczyk, "Face Detection and Tracking in a Video by Propagating Detection Probabilities," IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 25, no. 10, pp. 1215-1228, Oct. 2003.
[16] K. Toyama and G.D. Hager, "Incremental Focus of Attention for Robust Vision-Based Tracking," Int'l J. Computer Vision, vol. 35, no. 1, pp. 45-63, 1999.
[17] R. Sznitman and B. Jedynak, "Active Testing for Face Detection and Localization," IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 32, no. 10, pp. 1914-1920, Oct. 2010.
[18] S. Baker and I. Matthews, "Lucas-Kanade 20 Years On: A Unifying Framework Part 1: The Quantity Approximated, the Warp Update Rule, and the Gradient Descent Approximation," Int'l J. Computer Vision, vol. 56, no. 3, pp. 221-255, 2004.
[19] R. Richa, M. Balicki, E. Meisner, R. Sznitman, R.H. Taylor, and G.D. Hager, "Visual Tracking of Surgical Tools for Proximity Detection in Retinal Surgery," Information Processing in Computer Assisted Interventions, vol. 6689, pp. 55-66, 2011.
[20] R. Sznitman, H. Lin, M. Gupta, and G. Hager, "Active Background Modeling: Actors on a Stage," Proc. IEEE Int'l Conf. Computer Vision Workshops, pp. 1222-1228, 2009.
[21] D. Burschka, J. Corso, M. Dewan, W. Lau, H. Li, H. Lin, P. Marayong, N. Ramay, G. Hager, B. Hoffman, D. Larkin, and C. Hasser, "Navigating Inner Space: 3-D Assistance for Minimally Invasive Surgery," Robotics and Autonomous Systems, vol. 52, pp. 5-26, 2005.
[22] D. Uecker, C. Leem, Y. Wang, and Y. Wang, "Automated Instrument Tracking in Robotically Assisted Laparoscopic Surgery," J. Image Guided Surgery, vol. 22, no. 6, pp. 429-437, 1995.
[23] Y. Wang, D. Uecker, and Y. Wang, "A New Framework for Vision-Enabled and Robotically Assisted Minimally Invasive Surgery," Computerized Medical Imaging and Graphics, vol. 1, no. 6, pp. 308-325, 1998.
[24] C. Doignon, F. Nageotte, and M. de Mathelin, "Detection of Grey Regions in Color Images: Application to the Segmentation of a Surgical Instrument in Robotized Laparoscopy," Proc. IEEE/RSJ Int'l Conf. Intelligent Robots and Systems, vol. 4, pp. 3394-3399, 2004.
[25] S.J. McKenna, H. Nait-Charif, and T. Frank, "Towards Video Understanding for Laparoscopic Surgery: Instrument Tracking," Proc. Image and Vision Computing New Zealand Conf., 2005.
[26] J. Climent and P. Mares, "Automatic Instrument Localization in Laparoscopic Surgery," Electronic Letters on Computer Vision and Image Analysis, vol. 4, no. 1, pp. 21-31, 2004.
[27] S. Voros, J.A. Long, and P. Cinquin, "Automatic Detection of Instruments in Laparoscopic Images: A First Step towards High-Level Command of Robotic Endoscopic Holders," Int'l J. Robotic Research, vol. 26, nos. 11/12, pp. 1173-1190, 2007.
[28] A. Casals, J. Amat, and E. Laporte, "Automatic Guidance of an Assistant Robot in Laparoscopic Surgery," Proc. IEEE Int'l Conf. Robotics and Automation, pp. 895-900, 1996.
[29] G. Wei, K. Arbter, and G. Hirzinger, "Real-Time Visual Servoing for Laparoscopic Surgery. Controlling Robot Motion with Color Image Segmentation," IEEE Eng. in Medicine and Biology Magazine, vol. 16, no. 1, pp. 40-45, Jan./Feb. 1997.
[30] O. Tonet, T.U. Ramesh, G. Megali, and P. Dario, "Image Analysis-Based Approach for Localization of Endoscopic Tools," Proc. Surgetica Conf., pp. 221-228, 2005.
[31] Z. Pezzementi, S. Voros, and G. Hager, "Articulated Object Tracking by Rendering Consistent Appearance Parts," Proc. IEEE Int'l Conf. Robotics and Automation, pp. 3940-3947, 2009.
[32] M. Dewan, P. Marayong, A.M. Okamura, and G.D. Hager, "Vision-Based Assistance for Ophthalmic Microsurgery," Proc. Seventh Int'l Conf. Medical Image Computing and Computer Assisted Intervention, pp. 49-57, 2002.
[33] R. Sznitman, A. Basu, R. Richa, J. Handa, P. Gehlbach, R.T.H.B. Jedynak, and G.D. Hager, "Unified Detection and Tracking in Retinal Microsurgery," Proc. Int'l Conf. Medical Image Computing and Computer Assisted Intervention, pp. 1-8, 2011.
[34] S. Thrun, W. Burgard, and D. Fox, Probabilistic Robotics (Intelligent Robotics and Autonomous Agents). The MIT Press, 2005.
[35] G. Hager, Task-Directed Sensor Fusion and Planning: A Computational Approach. Springer, 1990.
[36] N.S. Peng, J. Yang, and Z. Liu, "Mean Shift Blob Tracking with Kernel Histogram Filtering and Hypothesis Testing," Pattern Recognition Letters, vol. 26, no. 5, pp. 605-614, 2005.
[37] T.M. Cover and J.A. Thomas, Elements of Information Theory. Wiley-Interscience, 1991.
[38] J. Pakkanen, J. Iivarinen, and E. Oja, "The Evolving Tree—A Novel Self-Organizing Network for Data Analysis," Neural Processing Letters, vol. 20, no. 3, pp. 199-211, 2004.
[39] R. Richa, R. Sznitman, R.H. Taylor, and G.D. Hager, "Visual Tracking Using the Sum of Conditional Variance," Proc. IEEE Conf. Intelligent Robots and Systems, pp. 2953-2958, 2011.
[40] P. Kazanzides, S. DiMaio, A. Deguet, B. Vagvolgyi, M. Balicki, C. Schneider, R. Kumar, A. Jog, B. Itkowitz, C. Hasser, and R.H. Taylor, "The Surgical Assistant Workstation (SAW) in Minimally Invasive Surgery and Microsurgery," Proc. Int'l Conf. Medical Image Computing and Systems and Assisted Interventions, 2010.
[41] R. Sznitman, D. Rother, J. Handa, P. Gehlbach, G.D. Hager, and R.H. Taylor, "Adaptive Multispectral Illumination for Retinal Microsurgery," Proc. Int'l Conf. Medical Image Computing and Computer Assisted Intervention, pp. 465-472, 2010.
41 ms
(Ver 2.0)

Marketing Automation Platform Marketing Automation Tool