Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Rethinking Early Stopping: Refine, Then Calibrate

Created by
  • Haebom

Author

Eugene Berta, David Holzm uller, Michael I. Jordan, Francis Bach

Outline

This paper uses an appropriate loss function such as cross entropy to assess the quality of probabilistic predictions of machine learning classifiers, and points out that it can be decomposed into two components: calibration error and refinement error. The calibration error evaluates the overall under/overconfidence, while the refinement error measures the ability to distinguish different classes. This paper presents a new variational formulation of the calibration-refinement decomposition, which provides a new perspective on post-calibration and allows fast estimation of different terms. Through this, we provide theoretical and experimental evidence that the calibration error and refinement error are not minimized simultaneously during training. Therefore, choosing the optimal epoch based on the validation loss leads to a suboptimal trade-off for both terms. To address this, this paper proposes a method (Refine... then Calibrate) that minimizes only the refinement error during training before minimizing the post-calibration error using standard techniques. This method is seamlessly integrated into all classifiers and consistently improves performance on a variety of classification tasks.

Takeaways, Limitations

Takeaways:
We point out the limitations of the existing validation loss-based optimal epoch selection method and propose a new approach that separates and optimizes the calibration error and refinement error.
We experimentally demonstrate that the "Refine... then Calibrate" strategy can improve performance on various classification tasks.
We provide a novel method for efficiently estimating correction and refinement errors via variational formulations.
Limitations:
Further research is needed on the generalization performance of the proposed method. It may overfit to certain datasets or classifiers.
The "Refine... then Calibrate" strategy may lack guidelines for setting optimization parameters at each step.
Extensive experimental results for different types of classifiers and loss functions are required.
👍