Facial asymmetry quantification for expression invariant human identification

作者:

Highlights:

摘要

We investigate facial asymmetry as a biometric under expression variation. For the first time, we have defined two types of quantified facial asymmetry measures that are easily computable from facial images and videos. Our findings show that the asymmetry measures of automatically selected facial regions capture individual differences that are relatively stable to facial expression variations. More importantly, a synergy is achieved by combining facial asymmetry information with conventional EigenFace and FisherFace methods. We have assessed the generality of these findings across two publicly available face databases: Using a random subset of 110 subjects from the FERET database, a 38% classification error reduction rate is obtained. Error reduction rates of 45–100% are achieved on 55 subjects from the Cohn–Kanade AU-Coded Facial Expression Database. These results suggest that facial asymmetry may provide complementary discriminative information to human identification methods, which has been missing in automatic human identification.

论文关键词:

论文评审过程:Received 15 February 2002, Accepted 24 March 2003, Available online 1 August 2003.

论文官网地址:https://doi.org/10.1016/S1077-3142(03)00078-X