DocumentCode :
178907
Title :
Watch-List Screening Using Ensembles Based on Multiple Face Representations
Author :
Bashbaghi, S. ; Granger, E. ; Sabourin, R. ; Bilodeau, G.-A.
Author_Institution :
Lab. d´Imagerie de Vision et d´Intell. Artificielle, Univ. du Quebec, Montreal, QC, Canada
fYear :
2014
fDate :
24-28 Aug. 2014
Firstpage :
4489
Lastpage :
4494
Abstract :
Still-to-video face recognition (FR) is an important function in watch list screening, where faces captured over a network of video surveillance cameras are matched against reference stills of target individuals. Recognizing faces in a watch list is a challenging problem in semi - and unconstrained surveillance environments due to the lack of control over capture and operational conditions, and to the limited number of reference stills. This paper provides a performance baseline and guidelines for ensemble-based systems using a single high-quality reference still per individual, as found in many watch list screening applications. In particular, modular systems are considered, where an ensemble of template matchers based on multiple face representations is assigned to each individual of interest. During enrollment, multiple feature extraction (FE) techniques are applied to patches isolated in the reference still to generate diverse face-part representations that are robust to various nuisance factors (e.g., illumination and pose) encountered in video surveillance. The selection of relevant feature subsets, decision thresholds, and fusion functions of ensembles are achieved using faces of non-target individuals selected from reference videos (forming a universal background model). During operations, a face tracker gradually regroups faces captured from different people appearing in a scene, while each user-specific ensemble generates a decision per face capture. This leads to robust spatio-temporal FR when accumulated ensemble predictions surpass a detection threshold. Simulation results obtained with the Chokepoint video dataset show a significant improvement to accuracy, (1) when performing score-level fusion of matchers, where patches-based and FE techniques generate ensemble diversity, (2) when defining feature subsets and decision thresholds for each individual matcher of an ensemble using non-target videos, and (3) when accumulating positive detections over mul- iple frames.
Keywords :
face recognition; feature extraction; image matching; image representation; video surveillance; chokepoint video dataset; decision thresholds; detection threshold; ensemble diversity; face tracker; feature extraction techniques; fusion functions; multiple FE techniques; multiple face representations; nontarget individuals; nuisance factors; patches-based techniques; performance baseline; reference videos; relevant feature subsets; robust spatiotemporal FR; semiconstrained surveillance environments; single high-quality reference; still-to-video face recognition; template matchers; unconstrained surveillance environments; universal background model; video surveillance cameras; watchlist screening applications; Cameras; Face; Feature extraction; Iron; Principal component analysis; Robustness; Video sequences;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Pattern Recognition (ICPR), 2014 22nd International Conference on
Conference_Location :
Stockholm
ISSN :
1051-4651
Type :
conf
DOI :
10.1109/ICPR.2014.768
Filename :
6977481
Link To Document :
بازگشت