Show simple item record

dc.contributor.authorZlatintsi, Athanasia
dc.contributor.authorKoutras, Petros
dc.contributor.authorMalandrakis, Nikolaos
dc.contributor.authorEfthymiou, Niki
dc.contributor.authorPastra, Katerina
dc.contributor.authorPotamianos, Alexandros
dc.contributor.authorMaragos, Petros
dc.contributor.authorEvangelopoulos, Georgios
dc.date.accessioned2018-02-22T19:24:10Z
dc.date.available2018-02-22T19:24:10Z
dc.date.issued2017-08
dc.identifier.issn1687-5281
dc.identifier.issn1687-5176
dc.identifier.urihttp://hdl.handle.net/1721.1/113872
dc.description.abstractAbstract: Research related to computational modeling for machine-based understanding requires ground truth data for training, content analysis, and evaluation. In this paper, we present a multimodal video database, namely COGNIMUSE, annotated with sensory and semantic saliency, events, cross-media semantics, and emotion. The purpose of this database is manifold; it can be used for training and evaluation of event detection and summarization algorithms, for classification and recognition of audio-visual and cross-media events, as well as for emotion tracking. In order to enable comparisons with other computational models, we propose state-of-the-art algorithms, specifically a unified energy-based audio-visual framework and a method for text saliency computation, for the detection of perceptually salient events from videos. Additionally, a movie summarization system for the automatic production of summaries is presented. Two kinds of evaluation were performed, an objective based on the saliency annotation of the database and an extensive qualitative human evaluation of the automatically produced summaries, where we investigated what composes high-quality movie summaries, where both methods verified the appropriateness of the proposed methods. The annotation of the database and the code for the summarization system can be found at http://cognimuse.cs.ntua.gr/database Keywords: Video database, SaliencyCross-media relations, Emotion annotation, Audio-visual events, Video summarization .en_US
dc.publisherSpringer International Publishingen_US
dc.relation.isversionofhttp://dx.doi.org/10.1186/s13640-017-0194-1en_US
dc.rightsCreative Commons Attributionen_US
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/en_US
dc.sourceSpringer International Publishingen_US
dc.titleCOGNIMUSE: a multimodal video database annotated with saliency, events, semantics and emotion with application to summarizationen_US
dc.typeArticleen_US
dc.identifier.citationZlatintsi, Athanasia, et al. “COGNIMUSE: A Multimodal Video Database Annotated with Saliency, Events, Semantics and Emotion with Application to Summarization.” EURASIP Journal on Image and Video Processing, vol. 2017, no. 1, Dec. 2017.en_US
dc.contributor.departmentMcGovern Institute for Brain Research at MITen_US
dc.contributor.mitauthorEvangelopoulos, Georgios
dc.relation.journalEURASIP Journal on Image and Video Processingen_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2017-08-08T04:02:46Z
dc.language.rfc3066en
dc.rights.holderThe Author(s)
dspace.orderedauthorsZlatintsi, Athanasia; Koutras, Petros; Evangelopoulos, Georgios; Malandrakis, Nikolaos; Efthymiou, Niki; Pastra, Katerina; Potamianos, Alexandros; Maragos, Petrosen_US
dspace.embargo.termsNen_US
dc.identifier.orcidhttps://orcid.org/0000-0003-2240-1801
mit.licensePUBLISHER_CCen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record