Monocular head pose estimation using generalized adaptive view-based appearance model

作者:

Highlights:

摘要

Accurately estimating the person’s head position and orientation is an important task for a wide range of applications such as driver awareness, meeting analysis and human-robot interaction. Over the past two decades, many approaches have been suggested to solve this problem, each with its own advantages and disadvantages. In this paper, we present a probabilistic framework called Generalized Adaptive View-based Appearance Model (GAVAM) which integrates the advantages from three of these approaches: (1) the automatic initialization and stability of static head pose estimation, (2) the relative precision and user-independence of differential registration, and (3) the robustness and bounded drift of keyframe tracking. In our experiments, we show how the GAVAM model can be used to estimate head position and orientation in real-time using a simple monocular camera. Our experiments on two previously published datasets show that the GAVAM framework can accurately track for a long period of time with an average accuracy of 3.5° and 0.75 in. when compared with an inertial sensor and a 3D magnetic sensor.

论文关键词:Head pose estimation,View-based appearance model,Keyframe tracking,Differential tracking,Rigid body tracking,Kalman filter update,Bounded drift

论文评审过程:Received 3 March 2009, Revised 29 July 2009, Accepted 2 August 2009, Available online 15 August 2009.

论文官网地址:https://doi.org/10.1016/j.imavis.2009.08.004