Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Minimizing Surrogate Losses for Decision-Focused Learning using Differentiable Optimization

Created by
  • Haebom

Author

Jayanta Mandi, Ali Irfan Mahmuto\u{g}ullar{\i}, Senne Berden, Tias Guns

Outline

This paper addresses the limitations of gradient-based DFL for optimization problems such as linear programming (LP) in decision-driven learning (DFL). Existing gradient-based DFL approaches attempt to address this issue in one of two ways: (a) smoothing the LP problem by adding quadratic regularization, thereby making it differentiable, or (b) minimizing a surrogate loss with information-rich (sub)gradients. However, this paper shows that approach (a) still suffers from the problem of zero gradients even when smoothed. Therefore, this paper proposes minimizing the surrogate loss even using differentiable optimization layers. Experimental results demonstrate that differentiable optimization layers achieve comparable or better regrets compared to existing surrogate-loss-based DFL methods through surrogate loss minimization. Specifically, we demonstrate that minimizing the surrogate loss using DYS-Net can achieve state-of-the-art regrets while significantly reducing training time.

Takeaways, Limitations

Takeaways:
Clarifies the limitations of gradient-based DFL for LP.
We demonstrate the effectiveness of surrogate loss minimization even when using differentiable optimization layers.
We demonstrate the feasibility of achieving state-of-the-art performance while reducing training time using DYS-Net.
Limitations:
Further research is needed to determine whether the proposed method in this paper can be applied to other types of optimization problems.
Further analysis is needed to determine how far DYS-Net's approximate solution differs from the optimal solution.
The generalizability of experimental results limited to specific types of LP problems needs to be examined.
👍