Now showing items 1-5 of 5
Seeing What You’re Told: Sentence-Guided Activity Recognition In Video
(Center for Brains, Minds and Machines (CBMM), arXiv, 2014-05-29)
We present a system that demonstrates how the compositional structure of events, in concert with the compositional structure of language, can interplay with the underlying focusing mechanisms in video action recognition, ...
The Compositional Nature of Event Representations in the Human Brain
(Center for Brains, Minds and Machines (CBMM), arXiv, 2014-07-14)
How does the human brain represent simple compositions of constituents: actors, verbs, objects, directions, and locations? Subjects viewed videos during neuroimaging (fMRI) sessions from which sentential descriptions of ...
Anchoring and Agreement in Syntactic Annotations
(Center for Brains, Minds and Machines (CBMM), arXiv, 2016-09-21)
Published in the Proceedings of EMNLP 2016 We present a study on two key characteristics of human syntactic annotations: anchoring and agreement. Anchoring is a well-known cognitive bias in human decision making, where ...
Seeing is Worse than Believing: Reading People’s Minds Better than Computer-Vision Methods Recognize Actions
We had human subjects perform a one-out-of-six class action recognition task from video stimuli while undergoing functional magnetic resonance imaging (fMRI). Support-vector machines (SVMs) were trained on the recovered ...
Do You See What I Mean? Visual Resolution of Linguistic Ambiguities
(Center for Brains, Minds and Machines (CBMM), arXiv, 2016-06-10)
Understanding language goes hand in hand with the ability to integrate complex contextual information obtained via perception. In this work, we present a novel task for grounded language understanding: disambiguating a ...