The Community for Technology Leaders
2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR) (2014)
Munich, Germany
Sept. 10, 2014 to Sept. 12, 2014
ISBN: 978-1-4799-6184-9
pp: 219-228
Taejin Ha , KAIST UVR Lab., 305-701, S. Korea
Steven Feiner , Columbia University, NY, NY 10027, USA
Woontack Woo , KAIST UVR Lab., 305-701, S. Korea
ABSTRACT
We introduce WeARHand, which allows a user to manipulate virtual 3D objects with a bare hand in a wearable augmented reality (AR) environment. Our method uses no environmentally tethered tracking devices and localizes a pair of near-range and far-range RGB-D cameras mounted on a head-worn display and a moving bare hand in 3D space by exploiting depth input data. Depth perception is enhanced through egocentric visual feedback, including a semi-transparent proxy hand. We implement a virtual hand interaction technique and feedback approaches, and evaluate their performance and usability. The proposed method can apply to many 3D interaction scenarios using hands in a wearable AR environment, such as AR information browsing, maintenance, design, and games.
INDEX TERMS
Cameras, Three-dimensional displays, Visualization, Image color analysis, Rendering (computer graphics), Image resolution, Equations,Virtual 3D Object Manipulation, Wearable Computing, Augmented Reality, 3D User Interfaces, Hand Interaction
CITATION
Taejin Ha, Steven Feiner, Woontack Woo, "WeARHand: Head-worn, RGB-D camera-based, bare-hand user interface with visually enhanced depth perception", 2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), vol. 00, no. , pp. 219-228, 2014, doi:10.1109/ISMAR.2014.6948431
191 ms
(Ver 3.3 (11022016))