Show simple item record

dc.contributor.authorWang, Zeyu
dc.contributor.authorShi, Yuanchun
dc.contributor.authorWang, Yuntao
dc.contributor.authorYao, Yuchen
dc.contributor.authorYan, Kun
dc.contributor.authorWang, Yuhan
dc.contributor.authorJi, Lei
dc.contributor.authorXu, Xuhai
dc.contributor.authorYu, Chun
dc.date.accessioned2024-06-06T16:47:57Z
dc.date.available2024-06-06T16:47:57Z
dc.date.issued2024-05-13
dc.identifier.issn2474-9567
dc.identifier.urihttps://hdl.handle.net/1721.1/155208
dc.description.abstractModern information querying systems are progressively incorporating multimodal inputs like vision and audio. However, the integration of gaze --- a modality deeply linked to user intent and increasingly accessible via gaze-tracking wearables --- remains underexplored. This paper introduces a novel gaze-facilitated information querying paradigm, named G-VOILA, which synergizes users' gaze, visual field, and voice-based natural language queries to facilitate a more intuitive querying process. In a user-enactment study involving 21 participants in 3 daily scenarios (p = 21, scene = 3), we revealed the ambiguity in users' query language and a gaze-voice coordination pattern in users' natural query behaviors with G-VOILA. Based on the quantitative and qualitative findings, we developed a design framework for the G-VOILA paradigm, which effectively integrates the gaze data with the in-situ querying context. Then we implemented a G-VOILA proof-of-concept using cutting-edge deep learning techniques. A follow-up user study (p = 16, scene = 2) demonstrates its effectiveness by achieving both higher objective score and subjective score, compared to a baseline without gaze data. We further conducted interviews and provided insights for future gaze-facilitated information querying systems.en_US
dc.publisherAssociation for Computing Machineryen_US
dc.relation.isversionof10.1145/3659623en_US
dc.rightsCreative Commons Attribution-Noncommercialen_US
dc.rights.urihttps://creativecommons.org/licenses/by-nc/4.0/en_US
dc.sourceAssociation for Computing Machineryen_US
dc.titleG-VOILA: Gaze-Facilitated Information Querying in Daily Scenariosen_US
dc.typeArticleen_US
dc.identifier.citationWang, Zeyu, Shi, Yuanchun, Wang, Yuntao, Yao, Yuchen, Yan, Kun et al. 2024. "G-VOILA: Gaze-Facilitated Information Querying in Daily Scenarios." Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 8 (2).
dc.relation.journalProceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologiesen_US
dc.identifier.mitlicensePUBLISHER_CC
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2024-06-01T07:58:52Z
dc.language.rfc3066en
dc.rights.holderThe author(s)
dspace.date.submission2024-06-01T07:58:53Z
mit.journal.volume8en_US
mit.journal.issue2en_US
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record