The Community for Technology Leaders
2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR) (2014)
Munich, Germany
Sept. 10, 2014 to Sept. 12, 2014
ISBN: 978-1-4799-6184-9
pp: 219-228
Taejin Ha , KAIST UVR Lab., 305-701, S. Korea
Steven Feiner , Columbia University, NY, NY 10027, USA
Woontack Woo , KAIST UVR Lab., 305-701, S. Korea
ABSTRACT
We introduce WeARHand, which allows a user to manipulate virtual 3D objects with a bare hand in a wearable augmented reality (AR) environment. Our method uses no environmentally tethered tracking devices and localizes a pair of near-range and far-range RGB-D cameras mounted on a head-worn display and a moving bare hand in 3D space by exploiting depth input data. Depth perception is enhanced through egocentric visual feedback, including a semi-transparent proxy hand. We implement a virtual hand interaction technique and feedback approaches, and evaluate their performance and usability. The proposed method can apply to many 3D interaction scenarios using hands in a wearable AR environment, such as AR information browsing, maintenance, design, and games.
INDEX TERMS
Cameras, Three-dimensional displays, Visualization, Image color analysis, Rendering (computer graphics), Image resolution, Equations
CITATION
Taejin Ha, Steven Feiner, Woontack Woo, "WeARHand: Head-worn, RGB-D camera-based, bare-hand user interface with visually enhanced depth perception", 2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), vol. 00, no. , pp. 219-228, 2014, doi:10.1109/ISMAR.2014.6948431
153 ms
(Ver 3.3 (11022016))