Show simple item record

dc.contributor.authorWang, Xiaogang
dc.contributor.authorMa, Xiaoxu
dc.contributor.authorGrimson, Eric
dc.date.accessioned2012-07-11T18:02:56Z
dc.date.available2012-07-11T18:02:56Z
dc.date.issued2008-04
dc.date.submitted2008-01
dc.identifier.issn0162-8828
dc.identifier.issn2160-9292
dc.identifier.urihttp://hdl.handle.net/1721.1/71587
dc.description.abstractWe propose a novel unsupervised learning framework to model activities and interactions in crowded and complicated scenes. Hierarchical Bayesian models are used to connect three elements in visual surveillance: low-level visual features, simple "atomic" activities, and interactions. Atomic activities are modeled as distributions over low-level visual features, and multi-agent interactions are modeled as distributions over atomic activities. These models are learnt in an unsupervised way. Given a long video sequence, moving pixels are clustered into different atomic activities and short video clips are clustered into different interactions. In this paper, we propose three hierarchical Bayesian models, Latent Dirichlet Allocation (LDA) mixture model, Hierarchical Dirichlet Process (HDP) mixture model, and Dual Hierarchical Dirichlet Processes (Dual-HDP) model. They advance existing language models, such as LDA [1] and HDP [2]. Our data sets are challenging video sequences from crowded traffic scenes and train station scenes with many kinds of activities co-occurring. Without tracking and human labeling effort, our framework completes many challenging visual surveillance tasks of board interest such as: (1) discovering typical atomic activities and interactions; (2) segmenting long video sequences into different interactions; (3) segmenting motions into different activities; (4) detecting abnormality; and (5) supporting high-level queries on activities and interactions.en_US
dc.description.sponsorshipUnited States. Defense Advanced Research Projects Agencyen_US
dc.description.sponsorshipSingapore. DSO National Laboratoriesen_US
dc.language.isoen_US
dc.publisherInstitute of Electrical and Electronics Engineers (IEEE)en_US
dc.relation.isversionofhttp://dx.doi.org/10.1109/tpami.2008.87en_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceIEEEen_US
dc.titleUnsupervised Activity Perception in Crowded and Complicated Scenes Using Hierarchical Bayesian Modelsen_US
dc.typeArticleen_US
dc.identifier.citationXiaogang Wang, Xiaoxu Ma, and W.E.L. Grimson. “Unsupervised Activity Perception in Crowded and Complicated Scenes Using Hierarchical Bayesian Models.” IEEE Transactions on Pattern Analysis and Machine Intelligence 31.3 (2009): 539–555. © Copyright 2009 IEEEen_US
dc.contributor.departmentMassachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratoryen_US
dc.contributor.approverGrimson, Eric
dc.contributor.mitauthorWang, Xiaogang
dc.contributor.mitauthorMa, Xiaoxu
dc.contributor.mitauthorGrimson, Eric
dc.relation.journalIEEE Transactions on Pattern Analysis and Machine Intelligenceen_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dspace.orderedauthorsXiaogang Wang; Xiaoxu Ma; Grimson, W.E.L.en
dc.identifier.orcidhttps://orcid.org/0000-0002-6192-2207
dspace.mitauthor.errortrue
mit.licensePUBLISHER_POLICYen_US
mit.metadata.statusComplete


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record