Search For:

Displaying 1-5 out of 5 total
Language model parameter estimation using user transcriptions
Found in: Acoustics, Speech, and Signal Processing, IEEE International Conference on
By Bo-June Hsu, James Glass
Issue Date:April 2009
pp. 4805-4808
In limited data domains, many effective language modeling techniques construct models with parameters to be estimated on an in-domain development set. However, in some domains, no such data exist beyond the unlabeled test corpus. In this work, we explore t...
 
On the phonetic information in ultrasonic microphone signals
Found in: Acoustics, Speech, and Signal Processing, IEEE International Conference on
By Karen Livescu, Bo Zhu, James Glass
Issue Date:April 2009
pp. 4621-4624
We study the phonetic information in the signal from an ultrasonic “microphone”, a device that emits an ultrasonic wave toward a speaker and receives the reflected, Doppler-shifted signal. This can be used in addition to audio to improve automatic speech r...
 
Multistream Articulatory Feature-Based Models for Visual Speech Recognition
Found in: IEEE Transactions on Pattern Analysis and Machine Intelligence
By Kate Saenko, Karen Livescu, James Glass, Trevor Darrell
Issue Date:September 2009
pp. 1700-1707
We study the problem of automatic visual speech recognition (VSR) using dynamic Bayesian network (DBN)-based models consisting of multiple sequences of hidden states, each corresponding to an articulatory feature (AF) such as lip opening (LO) or lip roundi...
 
Visual Speech Recognition with Loosely Synchronized Feature Streams
Found in: Computer Vision, IEEE International Conference on
By Kate Saenko, Karen Livescu, Michael Siracusa, Kevin Wilson,, James Glass, Trevor Darrell
Issue Date:October 2005
pp. 1424-1431
We present an approach to detecting and recognizing spoken isolated phrases based solely on visual input. We adopt an architecture that first employs discriminative detection of visual speech and articulatory features, and then performs recognition using a...
 
City browser: developing a conversational automotive HMI
Found in: Proceedings of the 27th international conference extended abstracts on Human factors in computing systems (CHI EA '09)
By Alexander Gruenstein, Bruce Mehler, Bryan Reimer, James Glass, Jarrod Orszulak, Jeff Zabel, Joseph Coughlin, Sean Liu, Shannon Roberts, Stephanie Seneff
Issue Date:April 2009
pp. 1-4
This paper introduces City Browser, a prototype multimodal, conversational, spoken language interface for automotive navigational aid and information access. A study designed to evaluate the impact of age and gender on device interaction errors, perception...
     
 1