Title :
Gender classification via lips: static and dynamic features
Author :
Stewart, Darryl ; Pass, Adrian ; Jianguo Zhang
Author_Institution :
Inst. of Electron., Commun. & Inf. Technol. (ECIT), Queens Univ. Belfast, Belfast, UK
Abstract :
Automatic gender classification has many security and commercial applications. Various modalities have been investigated for gender classification with face-based classification being the most popular. In some real-world scenarios the face may be partially occluded. In these circumstances a classification based on individual parts of the face known as local features must be adopted. The authors investigate gender classification using lip movements. They show for the first time that important gender-specific information can be obtained from the way in which a person moves their lips during speech. Furthermore, this study indicates that the lip dynamics during speech provide greater gender discriminative information than simply lip appearance. They also show that the lip dynamics and appearance contain complementary gender information such that a model which captures both traits gives the highest overall classification result. They use discrete cosine transform-based features and Gaussian mixture modelling to model lip appearance and dynamics and employ the XM2VTS database for their experiments. These experiments show that a model which captures lip dynamics along with appearance can improve gender classification rates by between 16 and 21% compared with models of only lip appearance.
Keywords :
Gaussian processes; discrete cosine transforms; feature extraction; gender issues; image classification; image motion analysis; Gaussian mixture modelling; XM2VTS database; automatic gender classification; complementary gender information; discrete cosine transform-based features; dynamic features; face-based classiflcation; gender discriminative information; gender-speciflc information; lip appearance modelling; lip dynamics modelling; lip movements; local features; static features;
Journal_Title :
Biometrics, IET
DOI :
10.1049/iet-bmt.2012.0021