Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

CMD-HAR: Cross-Modal Disentanglement for Wearable Human Activity Recognition

Created by
  • Haebom

Author

Hanyu Liu, Siyao Li, Ying Yu, Yixuan Jiang, Hang Xiao, Jingxi Long, Haotian Tang, Chao Li

Outline

This paper aims to address the issues of multimodal data mixing, activity heterogeneity, and complex model distribution in sensor-based human activity recognition (HAR). To this end, we propose a spatiotemporal attention modal decomposition alignment fusion strategy to address the mixed distribution problem of sensor data, capture key discriminative features of activities through multimodal spatiotemporal separate representations, and combine gradient modulation to mitigate data heterogeneity. In addition, we build a wearable deployment simulation system and demonstrate the effectiveness of the model through experiments using a number of public datasets.

Takeaways, Limitations

Takeaways:
A novel spatial-temporal attention-based fusion strategy for effectively processing multi-modal sensor data is presented.
Proposing a multi-modal separable representation method that effectively learns key discriminative features of activities
Applying gradient modulation techniques to alleviate data heterogeneity problems
Building a distribution simulation system considering the wearable environment
Validation of model performance through experiments using various public datasets
Limitations:
Further research is needed on the generalization performance of the proposed model.
Need for performance evaluation and verification in actual wearable environments
Need for performance evaluation of more diverse and complex activities
Energy efficiency and real-time processing performance must be taken into consideration
👍