DocumentCode
693761
Title
Comparing Image Representations for Training a Convolutional Neural Network to Classify Gender
Author
Choon-Boon Ng ; Yong-Haur Tay ; Bok-Min Goi
Author_Institution
Fac. of Eng. & Sci., Univ. Tunku Abdul Rahman, Kuala Lumpur, Malaysia
fYear
2013
fDate
3-5 Dec. 2013
Firstpage
29
Lastpage
33
Abstract
In this work, we evaluated the effect of different image representations on the classification performance of a convolutional neural network. Several different methods for normalization of the input data were also considered. The network was discriminatively trained for the task of gender classification of pedestrians. A publicly available dataset was used for training, containing both frontal and rear views of pedestrians. The best result was obtained using grayscale representation as compared to RGB and YUV, giving cross-validated accuracy of 81.5% on the dataset. The performance of the convolutional neural network is competitive and comparable to previous works on the same dataset.
Keywords
image classification; image colour analysis; image representation; neural nets; pedestrians; RGB; YUV; classification performance; convolutional neural network; cross-validated accuracy; frontal views; gender classification; grayscale representation; image representations; pedestrians; publicly available dataset; rear views; Computer architecture; Computer vision; Convolution; Gray-scale; Image color analysis; Neural networks; Training; convolutional neural network; gender classification; input representation; pedestrian;
fLanguage
English
Publisher
ieee
Conference_Titel
Artificial Intelligence, Modelling and Simulation (AIMS), 2013 1st International Conference on
Conference_Location
Kota Kinabalu
Print_ISBN
978-1-4799-3250-4
Type
conf
DOI
10.1109/AIMS.2013.13
Filename
6959890
Link To Document