Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Robust Multi-View Learning via Representation Fusion of Sample-Level Attention and Alignment of Simulated Perturbation

Created by
  • Haebom

Author

Jie Xu, Na Zhao, Gang Niu, Masashi Sugiyama, Xiaofeng Zhu

Outline

In this paper, we propose a robust MVL method (RML) that simultaneously performs representation fusion and alignment to overcome the limitations of multi-view learning (MVL) that integrates various types of data. RML uses a multi-view transformer fusion network to transform heterogeneous multi-view data into homogeneous word embeddings and obtains fused representations through a sample-level attention mechanism. In addition, we propose a multi-view contrastive learning framework that utilizes simulation-based perturbation to simulate incomplete data conditions, and aligns two fused representations obtained from noisy data and unavailable data through contrastive learning to learn discriminative and robust representations. RML is a self-supervised learning method and can be used as a plug-and-play module for multi-view unsupervised clustering, noisy label classification, and cross-modal hashing search. Experimental results verify the effectiveness of RML.

Takeaways, Limitations

Takeaways:
A novel MVL method for effectively integrating heterogeneous multi-view data is presented.
Robust representation learning for noisy and incomplete data
Self-supervised learning method, no need for separate label data
Applicable to various downstream tasks (unsupervised clustering, noisy label classification, cross-modal hashing search)
Available as a plug-and-play module
Limitations:
Lack of analysis of the computational cost and complexity of the proposed method.
Further validation is needed on generalization performance for various types of noise and incomplete data.
Potential bias exists for certain types of multi-view data
Possible lack of diversity in the experimental dataset (needs validation with additional diverse datasets)
👍