Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

C3T: Cross-modal Transfer Through Time for Sensor-based Human Activity Recognition

Created by
  • Haebom

Author

Abhi Kamboj, Anh Duy Nguyen, Minh N. Do

Outline

In this paper, we study a knowledge transfer method between time series modalities using a multimodal temporal representation space for human activity recognition (HAR) to leverage the potential of various sensors. In particular, we explore the setting where the modalities used for testing are unlabeled during training (Unsupervised Modality Adaptation (UMA). Existing UMA approaches are categorized into Student-Teacher or Contrastive Alignment methods, and point out the limitation that these methods generally compress continuous time data samples into a single latent vector, which hinders the ability to transfer temporal information through real-world temporal distortions. To address this, we propose Cross-modal Transfer Through Time (C3T), which preserves temporal information to better handle dynamic sensor data. C3T achieves this by aligning a series of temporal latent vectors across sensing modalities. Through extensive experiments on various camera+IMU datasets, we show that C3T outperforms existing UMA methods by at least 8% and exhibits excellent robustness against temporal distortions such as time shift, misalignment, and dilation. These results suggest that C3T has significant potential for developing generalizable models for time-series sensor data, opening up new avenues for various multi-modal applications.

Takeaways, Limitations

Takeaways:
Presentation of a new UMA method (C3T) that effectively utilizes temporal information from various sensor data
Suggesting the possibility of developing a multi-modal HAR model robust to time distortion
Demonstrated improved accuracy (at least 8% improvement) and robustness to time distortion compared to existing UMA methods
Contribute to the development of generalizable models for time-series sensor data
Limitations:
Further research is needed on the generalization performance of the proposed method.
Scalability verification for different types of sensors and activities is required.
Consideration should be given to additional temporal distortions that may occur when applying to real environments.
👍