Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Sufficient Invariant Learning for Distribution Shift

Created by
  • Haebom

Author

Taero Kim, Subeen Park, Sungjun Lim, Yonghan Jung, Krikamol Muandet, Kyungwoo Song

Outline

This paper addresses the problem of learning robust models under distributional shifts between training and test datasets. Existing invariant feature learning methods assume that invariant features are fully observed across both training and test sets, but in practice, this assumption is often violated. Therefore, models relying on invariant features absent from the test set may be less robust in new environments. To address this issue, this paper presents the Sufficient Invariant Learning (SIL) framework, a novel learning principle that learns a sufficient subset of invariant features instead of relying on a single feature. After demonstrating the limitations of existing invariant learning methods, we propose Adaptive Sharpness-Aware Group Distribution Robust Optimization (ASGDRO), a novel algorithm that learns diverse invariant features by finding common flat minima across environments. Theoretically, we demonstrate that finding common flat minima enables robust predictions based on diverse invariant features. Experimental evaluations on multiple datasets, including new benchmarks, confirm the robustness of ASGDRO to distributional shifts and highlight the limitations of existing methods.

Takeaways, Limitations

Takeaways:
We propose a Sufficiently Invariant Learning (SIL) framework that overcomes the limitations of existing invariant feature learning and enables more robust model learning.
We demonstrate that the ASGDRO algorithm can effectively learn various invariant features and improve robustness against distribution changes.
Theoretical support for finding a common flat minimum is crucial for robust prediction based on various invariant features.
Contributes to future research by providing a new benchmark dataset.
Limitations:
The computational complexity of the ASGDRO algorithm can be high.
Experimental evaluations are limited to specific datasets and tasks, requiring further research on generalizability.
More extensive research is needed on the applicability and limitations of the SIL framework.
👍