dc.contributor.author | Leifman, George | |
dc.contributor.author | Rudoy, Dmitry | |
dc.contributor.author | Swedish, Tristan | |
dc.contributor.author | Bayro-Corrochano, Eduardo | |
dc.contributor.author | Raskar, Ramesh | |
dc.date.accessioned | 2021-11-09T21:59:21Z | |
dc.date.available | 2021-11-09T21:59:21Z | |
dc.date.issued | 2017-10 | |
dc.identifier.uri | https://hdl.handle.net/1721.1/138091 | |
dc.description.abstract | © 2017 IEEE. In this paper we introduce a novel Depth-Aware Video Saliency approach to predict human focus of attention when viewing videos that contain a depth map (RGBD) on a 2D screen. Saliency estimation in this scenario is highly important since in the near future 3D video content will be easily acquired yet hard to display. Despite considerable progress in 3D display technologies, most are still expensive and require special glasses for viewing, so RGBD content is primarily viewed on 2D screens, removing the depth channel from the final viewing experience. We train a generative convolutional neural network that predicts the 2D viewing saliency map for a given frame using the RGBD pixel values and previous fixation estimates in the video. To evaluate the performance of our approach, we present a new comprehensive database of 2D viewing eye-fixation ground-truth for RGBD videos. Our experiments indicate that it is beneficial to integrate depth into video saliency estimates for content that is viewed on a 2D display. We demonstrate that our approach outperforms state-of-the-art methods for video saliency, achieving 15% relative improvement. | en_US |
dc.language.iso | en | |
dc.publisher | Institute of Electrical and Electronics Engineers (IEEE) | en_US |
dc.relation.isversionof | 10.1109/iccv.2017.188 | en_US |
dc.rights | Article is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use. | en_US |
dc.source | Computer Vision Foundation | en_US |
dc.title | Learning Gaze Transitions from Depth to Improve Video Saliency Estimation | en_US |
dc.type | Article | en_US |
dc.identifier.citation | Leifman, George, Rudoy, Dmitry, Swedish, Tristan, Bayro-Corrochano, Eduardo and Raskar, Ramesh. 2017. "Learning Gaze Transitions from Depth to Improve Video Saliency Estimation." | |
dc.contributor.department | Massachusetts Institute of Technology. Media Laboratory | en_US |
dc.eprint.version | Final published version | en_US |
dc.type.uri | http://purl.org/eprint/type/ConferencePaper | en_US |
eprint.status | http://purl.org/eprint/status/NonPeerReviewed | en_US |
dc.date.updated | 2019-08-02T12:59:41Z | |
dspace.date.submission | 2019-08-02T12:59:43Z | |
mit.metadata.status | Authority Work and Publication Information Needed | en_US |