The Community for Technology Leaders
2007 IEEE Conference on Computer Vision and Pattern Recognition (2007)
Minneapolis, MN, USA
June 17, 2007 to June 22, 2007
ISBN: 1-4244-1179-3
pp: 1-2
Dmitry N. Zotkin , Perceptual Interfaces and Reality Lab, Institute for Advanced Computer Studies (UMIACS), University
Vikas C. Raykar , Perceptual Interfaces and Reality Lab, Institute for Advanced Computer Studies (UMIACS), University
Ramani Duraiswami , Perceptual Interfaces and Reality Lab, Institute for Advanced Computer Studies (UMIACS), University
Larry S. Davis , Perceptual Interfaces and Reality Lab, Institute for Advanced Computer Studies (UMIACS), University
ABSTRACT
Many applications require the ability to track the 3-D motion of the subjects. We build a particle filter based framework for multimodal tracking using multiple cameras and multiple microphone arrays. In order to calibrate the resulting system, we propose a method to determine the locations of all microphones using at least five loudspeakers and under assumption that for each loudspeaker there exists a microphone very close to it. We derive the maximum likelihood (ML) estimator, which reduces to the solution of the non-linear least squares problem. We verify the correctness and robustness of the multimodal tracker and of the self-calibration algorithm both with Monte-Carlo simulations and on real data from three experimental setups.
INDEX TERMS
null
CITATION

L. S. Davis, V. C. Raykar, R. Duraiswami and D. N. Zotkin, "Multimodal Tracking for Smart Videoconferencing and Video Surveillance," 2007 IEEE Conference on Computer Vision and Pattern Recognition(CVPR), Minneapolis, MN, USA, 2007, pp. 1-2.
doi:10.1109/CVPR.2007.383525
95 ms
(Ver 3.3 (11022016))