Cross-Modal Federated Human Activity Recognition via Modality-Agnostic and Modality-Specific Representation Learning

Authors

  • Xiaoshan Yang National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence, University of Chinese Academy of Sciences Peng Cheng Laboratory
  • Baochen Xiong Zhengzhou University Peng Cheng Laboratory
  • Yi Huang National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence, University of Chinese Academy of Sciences
  • Changsheng Xu National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence, University of Chinese Academy of Sciences Peng Cheng Laboratory

DOI:

https://doi.org/10.1609/aaai.v36i3.20213

Keywords:

Computer Vision (CV), Machine Learning (ML)

Abstract

In this paper, we propose a new task of cross-modal federated human activity recognition (CMF-HAR), which is conducive to promote the large-scale use of the HAR model on more local devices. To address the new task, we propose a feature-disentangled activity recognition network (FDARN), which has five important modules of altruistic encoder, egocentric encoder, shared activity classifier, private activity classifier and modality discriminator. The altruistic encoder aims to collaboratively embed local instances on different clients into a modality-agnostic feature subspace. The egocentric encoder aims to produce modality-specific features that cannot be shared across clients with different modalities. The modality discriminator is used to adversarially guide the parameter learning of the altruistic and egocentric encoders. Through decentralized optimization with a spherical modality discriminative loss, our model can not only generalize well across different clients by leveraging the modality-agnostic features but also capture the modality-specific discriminative characteristics of each client. Extensive experiment results on four datasets demonstrate the effectiveness of our method.

Downloads

Published

2022-06-28

How to Cite

Yang, X., Xiong, B., Huang, Y., & Xu, C. (2022). Cross-Modal Federated Human Activity Recognition via Modality-Agnostic and Modality-Specific Representation Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 36(3), 3063-3071. https://doi.org/10.1609/aaai.v36i3.20213

Issue

Section

AAAI Technical Track on Computer Vision III