DocumentCode
702610
Title
Visual attention with deep neural networks
Author
Canziani, Alfredo ; Culurciello, Eugenio
Author_Institution
Weldon Sch. of Biomed. Eng., Purdue Univ., West Lafayette, IN, USA
fYear
2015
fDate
18-20 March 2015
Firstpage
1
Lastpage
3
Abstract
Animals use attentional mechanisms for being able to process enormous amount of sensory input in real time. Analogously, computerised systems could take advantage of similar techniques for achieving better timing performance. Visual attentional control uses bottom-up and top-down saliency maps for establishing the most relevant locations to observe. This article presents a novel fully-learnt unbiassed biologically plausible algorithm for computing both feature based and proto-object saliency maps, using a deep convolutional neural network simply trained on a single-class classification task, by unveiling its internal attentional apparatus. We are able to process 2 megapixels (MPs) colour images in real-time, i.e. at more than 10 frames per second, producing a 2MP map of interest.
Keywords
image classification; neural nets; bottom-up saliency maps; deep convolutional neural network; feature based saliency maps; fully-learnt unbiassed biologically plausible algorithm; internal attentional apparatus; proto-object saliency maps; single-class classification task; top-down saliency maps; visual attentional control; Biological neural networks; Computational modeling; Computer vision; Feature extraction; Real-time systems; Visualization;
fLanguage
English
Publisher
ieee
Conference_Titel
Information Sciences and Systems (CISS), 2015 49th Annual Conference on
Conference_Location
Baltimore, MD
Type
conf
DOI
10.1109/CISS.2015.7086900
Filename
7086900
Link To Document