The Community for Technology Leaders
Green Image
Issue No. 01 - January (2007 vol. 29)
ISSN: 0162-8828
pp: 65-81
An open vision problem is to automatically track the articulations of people from a video sequence. This problem is difficult because one needs to determine both the number of people in each frame and estimate their configurations. But, finding people and localizing their limbs is hard because people can move fast and unpredictably, can appear in a variety of poses and clothes, and are often surrounded by limb-like clutter. We develop a completely automatic system that works in two stages; it first builds a model of appearance of each person in a video and then it tracks by detecting those models in each frame ("tracking by model-building and detection”). We develop two algorithms that build models; one bottom-up approach groups together candidate body parts found throughout a sequence. We also describe a top-down approach that automatically builds people-models by detecting convenient key poses within a sequence. We finally show that building a discriminative model of appearance is quite helpful since it exploits structure in a background (without background-subtraction). We demonstrate the resulting tracker on hundreds of thousands of frames of unscripted indoor and outdoor activity, a feature-length film ("Run Lola Run”), and legacy sports footage (from the 2002 World Series and 1998 Winter Olympics). Experiments suggest that our system 1) can count distinct individuals, 2) can identify and track them, 3) can recover when it loses track, for example, if individuals are occluded or briefly leave the view, 4) can identify body configuration accurately, and 5) is not dependent on particular models of human motion.
People tracking, motion capture, surveillance.
Deva Ramanan, David A. Forsyth, Andrew Zisserman, "Tracking People by Learning Their Appearance", IEEE Transactions on Pattern Analysis & Machine Intelligence, vol. 29, no. , pp. 65-81, January 2007, doi:10.1109/TPAMI.2007.22
164 ms
(Ver 3.1 (10032016))