Facial Action Unit Intensity Estimation via Semantic Correspondence Learning with Dynamic Graph Convolution

Authors

  • Yingruo Fan The University of Hong Kong
  • Jacqueline Lam The University of Hong Kong
  • Victor Li The University of Hong Kong

DOI:

https://doi.org/10.1609/aaai.v34i07.6963

Abstract

The intensity estimation of facial action units (AUs) is challenging due to subtle changes in the person's facial appearance. Previous approaches mainly rely on probabilistic models or predefined rules for modeling co-occurrence relationships among AUs, leading to limited generalization. In contrast, we present a new learning framework that automatically learns the latent relationships of AUs via establishing semantic correspondences between feature maps. In the heatmap regression-based network, feature maps preserve rich semantic information associated with AU intensities and locations. Moreover, the AU co-occurring pattern can be reflected by activating a set of feature channels, where each channel encodes a specific visual pattern of AU. This motivates us to model the correlation among feature channels, which implicitly represents the co-occurrence relationship of AU intensity levels. Specifically, we introduce a semantic correspondence convolution (SCC) module to dynamically compute the correspondences from deep and low resolution feature maps, and thus enhancing the discriminability of features. The experimental results demonstrate the effectiveness and the superior performance of our method on two benchmark datasets.

Downloads

Published

2020-04-03

How to Cite

Fan, Y., Lam, J., & Li, V. (2020). Facial Action Unit Intensity Estimation via Semantic Correspondence Learning with Dynamic Graph Convolution. Proceedings of the AAAI Conference on Artificial Intelligence, 34(07), 12701-12708. https://doi.org/10.1609/aaai.v34i07.6963

Issue

Section

AAAI Technical Track: Vision