Show simple item record

dc.contributor.authorNarain, J
dc.contributor.authorJohnson, KT
dc.contributor.authorFerguson, C
dc.contributor.authorO'Brien, A
dc.contributor.authorTalkar, T
dc.contributor.authorWeninger, YZ
dc.contributor.authorWofford, P
dc.contributor.authorQuatieri, T
dc.contributor.authorPicard, Rosalind W.
dc.contributor.authorMaes, P
dc.date.accessioned2021-11-02T14:17:52Z
dc.date.available2021-11-02T14:17:52Z
dc.date.issued2020
dc.identifier.urihttps://hdl.handle.net/1721.1/137088
dc.description.abstract© 2020 Owner/Author. Nonverbal vocalizations contain important affective and communicative information, especially for those who do not use traditional speech, including individuals who have autism and are non- or minimally verbal (nv/mv). Although these vocalizations are often understood by those who know them well, they can be challenging to understand for the community-at-large. This work presents (1) a methodology for collecting spontaneous vocalizations from nv/mv individuals in natural environments, with no researcher present, and personalized in-the-moment labels from a family member; (2) speaker-dependent classification of these real-world sounds for three nv/mv individuals; and (3) an interactive application to translate the nonverbal vocalizations in real time. Using support-vector machine and random forest models, we achieved speaker-dependent unweighted average recalls (UARs) of 0.75, 0.53, and 0.79 for the three individuals, respectively, with each model discriminating between 5 nonverbal vocalization classes. We also present first results for real-time binary classification of positive- and negative-affect nonverbal vocalizations, trained using a commercial wearable microphone and tested in real time using a smartphone. This work informs personalized machine learning methods for non-traditional communicators and advances real-world interactive augmentative technology for an underserved population.en_US
dc.language.isoen
dc.publisherACMen_US
dc.relation.isversionof10.1145/3382507.3418854en_US
dc.rightsCreative Commons Attribution 4.0 International licenseen_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/en_US
dc.sourceACMen_US
dc.titlePersonalized Modeling of Real-World Vocalizations from Nonverbal Individualsen_US
dc.typeArticleen_US
dc.identifier.citationNarain, J, Johnson, KT, Ferguson, C, O'Brien, A, Talkar, T et al. 2020. "Personalized Modeling of Real-World Vocalizations from Nonverbal Individuals." ICMI 2020 - Proceedings of the 2020 International Conference on Multimodal Interaction.
dc.contributor.departmentMassachusetts Institute of Technology. Media Laboratory
dc.contributor.departmentLincoln Laboratory
dc.relation.journalICMI 2020 - Proceedings of the 2020 International Conference on Multimodal Interactionen_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2021-06-30T18:18:52Z
dspace.orderedauthorsNarain, J; Johnson, KT; Ferguson, C; O'Brien, A; Talkar, T; Weninger, YZ; Wofford, P; Quatieri, T; Picard, R; Maes, Pen_US
dspace.date.submission2021-06-30T18:18:55Z
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record