Loading…
Robust, accurate and efficient face recognition from a single training image: A uniform pursuit approach
Current face recognition techniques rely heavily on the large size and representativeness of the training sets, and most methods suffer degraded performance or fail to work if there is only one training sample per person available. This so-called “one sample problem” is a challenging issue in face r...
Saved in:
Published in: | Pattern recognition 2010-05, Vol.43 (5), p.1748-1762 |
---|---|
Main Authors: | , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Current face recognition techniques rely heavily on the large size and representativeness of the training sets, and most methods suffer degraded performance or fail to work if there is only one training sample per person available. This so-called “one sample problem” is a challenging issue in face recognition. In this paper, we propose a novel feature extraction method named uniform pursuit to address the one sample problem. The underlying idea is that most recognition errors are due to the confusions between faces that look very similar, and thus one can reduce the risk of recognition error by mapping the close class prototypes to be distant, i.e., uniforming the pairwise distances between different class prototypes. Specifically, the UP method pursues, in the whitened PCA space, the low dimensional projections that reduce the local confusion between the similar faces. The resulting low dimensional transformed features are robust against the complex image variations such as those caused by lighting and aging. A standardized procedure on the large-scale FERET and FRGC databases is applied to evaluate the one sample problem. Experimental results show that the robustness, accuracy and efficiency of the proposed UP method compare favorably to the state-of-the-art one sample based methods. |
---|---|
ISSN: | 0031-3203 1873-5142 |
DOI: | 10.1016/j.patcog.2009.12.004 |