Gaze from Origin: Learning for Generalized Gaze Estimation by Embedding the Gaze Frontalization Process
DOI:
https://doi.org/10.1609/aaai.v38i6.28452Keywords:
CV: Biometrics, Face, Gesture & Pose, ML: Transfer, Domain Adaptation, Multi-Task LearningAbstract
Gaze estimation aims to accurately estimate the direction or position at which a person is looking. With the development of deep learning techniques, a number of gaze estimation methods have been proposed and achieved state-of-the-art performance. However, these methods are limited to within-dataset settings, whose performance drops when tested on unseen datasets. We argue that this is caused by infinite and continuous gaze labels. To alleviate this problem, we propose using gaze frontalization as an auxiliary task to constrain gaze estimation. Based on this, we propose a novel gaze domain generalization framework named Gaze Frontalization-based Auxiliary Learning (GFAL) Framework which embeds the gaze frontalization process, i.e., guiding the feature so that the eyeball can rotate and look at the front (camera), without any target domain information during training. Experimental results show that our proposed framework is able to achieve state-of-the-art performance on gaze domain generalization task, which is competitive with or even superior to the SOTA gaze unsupervised domain adaptation methods.Downloads
Published
2024-03-24
How to Cite
Xu, M., & Lu, F. (2024). Gaze from Origin: Learning for Generalized Gaze Estimation by Embedding the Gaze Frontalization Process. Proceedings of the AAAI Conference on Artificial Intelligence, 38(6), 6333-6341. https://doi.org/10.1609/aaai.v38i6.28452
Issue
Section
AAAI Technical Track on Computer Vision V