Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Evaluating the Efficacy of LLM-Based Reasoning for Multiobjective HPC Job Scheduling

Created by
  • Haebom

Author

Prachi Jadhav, Hongwei Jin, Ewa Deelman, Prasanna Balaprakash

Outline

This paper proposes a large-scale language model (LLM)-based scheduler based on the ReAct framework to address high-performance computing (HPC) task scheduling problems. Existing heuristic-based methods (FCFS, SJF) and optimization techniques lack adaptability to dynamic workloads and cannot simultaneously optimize multiple objectives. The proposed LLM-based scheduler utilizes scratchpad memory to track scheduling history, improves decision-making through natural language feedback, and ensures feasibility and safety through a constraint enforcement module. Evaluations on various real-world HPC workload scenarios using OpenAI's O4-Mini and Anthropic's Claude 3.7 demonstrate that the LLM-based scheduler effectively balances multiple objectives and provides transparent inference through natural language tracking. It exhibits excellent constraint satisfaction and adapts to diverse workloads without domain-specific learning. However, the tradeoff between inference quality and computational overhead remains a challenge for real-time deployment. This paper is the first comprehensive study of the application of inferential LLM to HPC scheduling, demonstrating its potential for multi-objective optimization while highlighting the limitations of computational efficiency.

Takeaways, Limitations

Takeaways:
We demonstrate that LLM-based scheduling can effectively balance multiple objectives (minimizing makespan, reducing latency, optimizing resource utilization, and ensuring fairness) across diverse HPC workloads.
Natural language tracking can help ensure transparency in the scheduling process and make decision-making easier to understand.
It provides the flexibility to adapt to diverse workloads without domain-specific learning.
We present the potential application of LLM to complex scheduling problems.
Limitations:
The tradeoff between LLM's inference quality and computational overhead can make it challenging for real-time deployment.
The computational cost of LLM is significant, requiring an economic efficiency review for application in real HPC environments.
👍