Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Complex Model Transformations by Reinforcement Learning with Uncertain Human Guidance

Created by
  • Haebom

Author

Kyanna Dagenais, Istvan David

Outline

This paper presents a reinforcement learning (RL)-based framework for efficiently developing complex model transformation (MT) sequences in model-based engineering. Complex MT sequences are required for a variety of problems, including model synchronization, automatic model recovery, and design space exploration. However, manually developing them is error-prone and challenging. In this paper, we propose an approach and technical framework that enables an RL agent to find optimal MT sequences using user advice, which may include uncertainty. We map user-defined MTs to RL primitives and execute them as RL programs to find optimal MT sequences. Experimental results demonstrate that even under uncertainty, user advice significantly improves RL performance, contributing to more efficient development of complex MTs. This study advances RL-based human-in-the-loop engineering methodology by addressing the tradeoff between the certainty and timing of user advice.

Takeaways, Limitations

Takeaways:
We demonstrate that an RL-based framework that incorporates uncertain user advice can improve the efficiency of developing complex model transformation sequences.
We present a novel approach to effectively apply RL to human-in-the-loop engineering methodologies.
We provide insights into the design of practical RL-based systems by considering the trade-off between certainty and timing of user advice.
Limitations:
Further research is needed to determine the generality and scalability of the proposed framework. Its applicability to various types of models and problems needs to be further verified.
A more in-depth analysis of performance changes based on the quality and quantity of user advice is needed.
Further research is needed to determine its applicability and practicality in real engineering environments.
👍