This page organizes papers related to artificial intelligence published around the world. This page is summarized using Google Gemini and is operated on a non-profit basis. The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.
PRISM: Self-Pruning Intrinsic Selection Method for Training-Free Multimodal Data Selection
Created by
Haebom
Author
Jinhe Bi, Yifan Wang, Danqi Yan, Aniri, Wenke Huang, Zengjie Jin, Xiaowen Ma, Artur Hecker, Mang Ye, Xun Xiao, Hinrich Schuetze, Volker Tresp, Yunpu Ma
Outline
This paper proposes PRISM, a training-free framework, to address the computational overhead inherent in data selection for visual guidance tuning. We identify anisotropy in visual feature distributions as a key factor contributing to the inefficiency of existing methods. To address this, PRISM models intrinsic visual semantics to eliminate global semantic drift. Experimental results demonstrate that PRISM reduces data selection and model tuning time by 30%, while outperforming existing models across multiple benchmarks.
Takeaways, Limitations
•
Takeaways:
◦
Significantly improves computational efficiency through PRISM, a data selection framework that requires no training.
◦
We present a new problem of anisotropy in the distribution of visual features and propose an approach to solve it.
◦
Shows improved performance compared to existing models in various benchmarks.
•
Limitations:
◦
The paper does not specifically mention Limitations (however, further research may be needed to determine generalizability across datasets and applicability to other MLLM architectures).