Facial expression recognition from near-infrared videos

作者:

Highlights:

摘要

Facial expression recognition is to determine the emotional state of the face regardless of its identity. Most of the existing datasets for facial expressions are captured in a visible light spectrum. However, the visible light (VIS) can change with time and location, causing significant variations in appearance and texture. In this paper, we present a novel research on a dynamic facial expression recognition, using near-infrared (NIR) video sequences and LBP-TOP (Local binary patterns from three orthogonal planes) feature descriptors. NIR imaging combined with LBP-TOP features provide an illumination invariant description of face video sequences. Appearance and motion features in slices are used for expression classification, and for this, discriminative weights are learned from training examples. Furthermore, component-based facial features are presented to combine geometric and appearance information, providing an effective way for representing the facial expressions. Experimental results of facial expression recognition using a novel Oulu-CASIA NIR&VIS facial expression database, a support vector machine and sparse representation classifiers show good and robust results against illumination variations. This provides a baseline for future research on NIR-based facial expression recognition.

论文关键词:Facial expression recognition,Spatiotemporal descriptors,Near-infrared (NIR),Visible light (VIS),Component-based facial features

论文评审过程:Received 22 October 2010, Revised 14 June 2011, Accepted 1 July 2011, Available online 8 July 2011.

论文官网地址:https://doi.org/10.1016/j.imavis.2011.07.002