An integrated model of visual attention using shape-based features
Author(s)
Poggio, Tomaso; Serre, Thomas; Tan, Cheston; Chikkerur, Sharat
DownloadMIT-CSAIL-TR-2009-029.pdf (15.47Mb)
Additional downloads
Other Contributors
Center for Biological and Computational Learning (CBCL)
Advisor
Tomaso Poggio
Metadata
Show full item recordAbstract
Apart from helping shed some light on human perceptual mechanisms, modeling visual attention has important applications in computer vision. It has been shown to be useful in priming object detection, pruning interest points, quantifying visual clutter as well as predicting human eye movements. Prior work has either relied on purely bottom-up approaches or top-down schemes using simple low-level features. In this paper, we outline a top-down visual attention model based on shape-based features. The same shape-based representation is used to represent both the objects and the scenes that contain them. The spatial priors imposed by the scene and the feature priors imposed by the target object are combined in a Bayesian framework to generate a task-dependent saliency map. We show that our approach can predict the location of objects as well as match eye movements (92% overlap with human observers). We also show that the proposed approach performs better than existing bottom-up and top-down computational models.
Date issued
2009-06-20Series/Report no.
CBCL-278MIT-CSAIL-TR-2009-029
Keywords
attention, bayesian network