Loading…

Regression-based intensity estimation of facial action units

Facial Action Coding System (FACS) is the de facto standard in the analysis of facial expressions. FACS describes expressions in terms of the configuration and strength of atomic units called Action Units: AUs. FACS defines 44 AUs and each AU intensity is defined on a nonlinear scale of five grades....

Full description

Saved in:
Bibliographic Details
Published in:Image and vision computing 2012-10, Vol.30 (10), p.774-784
Main Authors: Savran, Arman, Sankur, Bulent, Taha Bilge, M.
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Facial Action Coding System (FACS) is the de facto standard in the analysis of facial expressions. FACS describes expressions in terms of the configuration and strength of atomic units called Action Units: AUs. FACS defines 44 AUs and each AU intensity is defined on a nonlinear scale of five grades. There has been significant progress in the literature on the detection of AUs. However, the companion problem of estimating the AU strengths has not been much investigated. In this work we propose a novel AU intensity estimation scheme applied to 2D luminance and/or 3D surface geometry images. Our scheme is based on regression of selected image features. These features are either non-specific, that is, those inherited from the AU detection algorithm, or are specific in that they are selected for the sole purpose of intensity estimation. For thoroughness, various types of local 3D shape indicators have been considered, such as mean curvature, Gaussian curvature, shape index and curvedness, as well as their fusion. The feature selection from the initial plethora of Gabor moments is instrumented via a regression that optimizes the AU intensity predictions. Our AU intensity estimator is person-independent and when tested on 25 AUs that appear singly or in various combinations, it performs significantly better than the state-of-the-art method which is based on the margins of SVMs designed for AU detection. When evaluated comparatively, one can see that the 2D and 3D modalities have relative merits per upper face and lower face AUs, respectively, and that there is an overall improvement if 2D and 3D intensity estimations are used in fusion. ► Regression for person-independent estimation of facial action unit intensities ► Mean curvature, Gaussian curvature, shape index, curvedness for 3D estimation ► Fusing curvature, Gaussian curvature, and curvedness achieves the best 3D estimation ► SVM regression of AdaBoost/AdaBoost.RT selected features is superior to SVM margins ► Modality fusion overcomes deficiencies of 3D in upper face and 2D in lower face
ISSN:0262-8856
1872-8138
DOI:10.1016/j.imavis.2011.11.008