From Video to Images: Contrastive Pretraining for Emotion Recognition from Single Image (Student Abstract)

Authors

  • Bhanu Garg University of California San Diego
  • Kijun Kim University of California San Diego
  • Sudhanshu Ranjan University of California San Diego

DOI:

https://doi.org/10.1609/aaai.v36i11.21612

Keywords:

Self Supervised Learning, Contrastive Learning, Emotion Recognition, Computer Vision, Deep Learning

Abstract

Emotion detection from face is an important problem and has received attention from industry and academia. Although emotion recognition from videos has a very high performance, emotion recognition from a single image stays a challenging task. In this paper, we try to use information from videos to do emotion recognition on a single image. More specifically, we leverage contrastive loss for pretraining the network on the videos and experiment with different sampling methods to select consistently hard triplets for continual learning of the network. Once the embeddings have been trained, we test them on a standard emotion classification task. Our method significantly improves the performance of the models and shows the efficacy of self-supervision in emotion recognition.

Downloads

Published

2022-06-28

How to Cite

Garg, B., Kim, K., & Ranjan, S. (2022). From Video to Images: Contrastive Pretraining for Emotion Recognition from Single Image (Student Abstract). Proceedings of the AAAI Conference on Artificial Intelligence, 36(11), 12951-12952. https://doi.org/10.1609/aaai.v36i11.21612