DocumentCode
730830
Title
Probabilistic features for connecting eye gaze to spoken language understanding
Author
Prokofieva, Anna ; Slaney, Malcolm ; Hakkani-Tur, Dilek
Author_Institution
Microsoft Res., Mountain View, CA, USA
fYear
2015
fDate
19-24 April 2015
Firstpage
5311
Lastpage
5315
Abstract
Many users obtain content from a screen and want to make requests of a system based on items that they have seen. Eye-gaze information is a valuable signal in speech recognition and spoken-language understanding (SLU) because it provides context for a user´s next utterance-what the user says next is probably conditioned on what they have seen. This paper investigates three types of features for connecting eye-gaze information to an SLU system: lexical, and two types of eye-gaze features. These features help us to understand which object (i.e. a link) that a user is referring to on a screen. We show a 17% absolute performance improvement in the referenced-object F-score by adding eye-gaze features to conventional methods based on a lexical comparison of the spoken utterance and the text on the screen.
Keywords
gaze tracking; natural language processing; probability; speech recognition; user interfaces; eye gaze features; eye-gaze information; lexical feature; next utterance context; probabilistic features; referenced object F-score; speech recognition; spoken language understanding; Conferences; Face; Heating; Probabilistic logic; Speech; Speech recognition; Standards; Spoken language understanding; classification; eye gaze; heat maps; referring expression resolution;
fLanguage
English
Publisher
ieee
Conference_Titel
Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
Conference_Location
South Brisbane, QLD
Type
conf
DOI
10.1109/ICASSP.2015.7178985
Filename
7178985
Link To Document