Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Irredundant $k$-Fold Cross-Validation

Created by
  • Haebom

Author

Jesus S. Aguilar-Ruiz

Outline

Conventional k-fold cross-validation suffers from the problem of using each instance for training (k-1) times and testing once, resulting in redundancy and a disproportionate impact on the learning process due to multiple instances. In this paper, we present a novel method, Irredundant k-fold cross-validation, which ensures that each instance is used for both training and testing exactly once throughout the validation process. This ensures balanced dataset utilization, mitigates overfitting due to instance repetition, and enables more distinct differences in model analysis. Experimental results demonstrate that this method maintains hierarchical and model-independent performance across diverse datasets, while providing lower variance estimates and significantly reducing overall computational costs due to non-overlapping training partitions.

Takeaways, Limitations

Takeaways:
Ensures balanced utilization of the dataset.
Mitigates overfitting due to instance repetition.
Provides clearer differences in model comparison analysis.
It lowers the variance estimate while maintaining similar performance to the traditional k-fold cross-validation.
Significantly reduces overall computational costs.
Maintains hierarchical structure and is model-independent.
Limitations:
This paper does not specifically mention Limitations. Further experiments and analyses are needed to verify its performance in a wider range of situations.
👍