Title :
Computing 3D saliency from a 2D image
Author :
Ramenahalli, Sudarshan ; Niebur, Ernst
Author_Institution :
Dept. of Elec & Comput. Eng, Johns Hopkins Univ. Baltimore, Baltimore, MD, USA
Abstract :
A saliency map is a model of visual selective attention using purely bottom-up features of an image like color, intensity and orientation. Another bottom-up feature of visual input is depth, the distance between eye (or sensor) and objects in the visual field. In this report we study the effect of depth on saliency. Different from previous work, we do not use measured depth (disparity) information but, instead, compute a 3D depth map from the 2D image using a depth learning algorithm. This computed depth is then added as an additional feature channel to the 2D saliency map, and all feature channels are linearly combined with equal weights to obtain a 3-dimensional saliency map. We compare the efficacy of saliency maps (2D and 3D) in predicting human eye fixations using three different performance measures. The 3D saliency map outperforms its 2D counterpart in predicting human eye fixations on all measures. Perhaps surprisingly, our 3D saliency map computed from a 2D image performs better than an existing 3D saliency model that uses explicit depth information.
Keywords :
stereo image processing; 2D image; 3D depth map; 3D saliency computation; bottom-up feature; depth learning algorithm; image like color; image like intensity; image like orientation; three dimensional saliency map; visual selective attention; Accuracy; Robots;
Conference_Titel :
Information Sciences and Systems (CISS), 2013 47th Annual Conference on
Conference_Location :
Baltimore, MD
Print_ISBN :
978-1-4673-5237-6
Electronic_ISBN :
978-1-4673-5238-3
DOI :
10.1109/CISS.2013.6552297