Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Learning Hierarchical Domain Models Through Environment-Grounded Interaction

Created by
  • Haebom

Author

Claudius Kienle, Benjamin Alt, Oleg Arenz, Jan Peters

Outline

LODGE studies domain models that help autonomous agents generate interpretable plans to solve long-term tasks. Because a single, general domain model cannot handle diverse tasks in an open environment, agents must generate task-specific models on the fly. While LLM can generate such domains, its high error rate limits its applicability. LODGE is a framework for LLM and environment-based unsupervised domain learning. It leverages hierarchical abstraction and automated simulation to identify and correct mismatches between abstraction layers and between models and environments. LODGE is task-agnostic and generates predicates, operators, preconditions, and effects based solely on access to a simulator and a set of general, executable, low-level techniques. Experimental results demonstrate that LODGE produces more accurate domain models and higher task success rates than existing methods, while requiring less environmental interaction and no human feedback or demonstration.

Takeaways, Limitations

Takeaways:
We present an autonomous domain learning framework leveraging LLM and environments without human feedback or prior knowledge for autonomous open environment deployment.
Effectively identify and correct mismatches between models and environments through hierarchical abstraction and automatic simulation.
Generate generalized predicates, operators, etc. in a task-independent manner, so that they can be applied to various domains.
Experiments have proven improved accuracy and success rate compared to existing methods.
Limitations:
Explicit Limitations is not directly mentioned in the paper.
Since it depends on the generating ability of LLM, the possibility of errors in LLM itself cannot be completely ruled out.
Requires access to simulators and low-level technology sets, and may have dependencies on these elements.
Performance in complex environments other than the presented experimental domain requires further validation.
👍