Gaze from Origin: Learning for Generalized Gaze Estimation by Embedding the Gaze Frontalization Process

Authors

  • Mingjie Xu State Key Laboratory of VR Technology and Systems, School of CSE, Beihang University, Beijing, China
  • Feng Lu State Key Laboratory of VR Technology and Systems, School of CSE, Beihang University, Beijing, China Peng Cheng Laboratory, Shenzhen, China

DOI:

https://doi.org/10.1609/aaai.v38i6.28452

Keywords:

CV: Biometrics, Face, Gesture & Pose, ML: Transfer, Domain Adaptation, Multi-Task Learning

Abstract

Gaze estimation aims to accurately estimate the direction or position at which a person is looking. With the development of deep learning techniques, a number of gaze estimation methods have been proposed and achieved state-of-the-art performance. However, these methods are limited to within-dataset settings, whose performance drops when tested on unseen datasets. We argue that this is caused by infinite and continuous gaze labels. To alleviate this problem, we propose using gaze frontalization as an auxiliary task to constrain gaze estimation. Based on this, we propose a novel gaze domain generalization framework named Gaze Frontalization-based Auxiliary Learning (GFAL) Framework which embeds the gaze frontalization process, i.e., guiding the feature so that the eyeball can rotate and look at the front (camera), without any target domain information during training. Experimental results show that our proposed framework is able to achieve state-of-the-art performance on gaze domain generalization task, which is competitive with or even superior to the SOTA gaze unsupervised domain adaptation methods.

Downloads

Published

2024-03-24

How to Cite

Xu, M., & Lu, F. (2024). Gaze from Origin: Learning for Generalized Gaze Estimation by Embedding the Gaze Frontalization Process. Proceedings of the AAAI Conference on Artificial Intelligence, 38(6), 6333-6341. https://doi.org/10.1609/aaai.v38i6.28452

Issue

Section

AAAI Technical Track on Computer Vision V