Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Scalable Option Learning in High-Throughput Environments

Created by
  • Haebom

Author

Mikael Henaff, Scott Fujimoto, Michael Matthews, Michael Rabbit

Outline

Hierarchical reinforcement learning (RL) has the potential to enable effective decision-making over long periods of time. Existing approaches, while promising, have yet to realize the benefits of large-scale training. This study identifies and addresses several key challenges in scaling online hierarchical RL to high-throughput environments. We propose Scalable Option Learning (SOL), a highly scalable hierarchical RL algorithm that achieves ~35x higher throughput than existing hierarchical approaches. We demonstrate the performance and scalability of SOL by training a hierarchical agent using 30 billion frames of experience in the complex game NetHack, significantly outperforming flat agents and demonstrating positive scaling trends. We also validate SOL in MiniHack and Mujoco environments, demonstrating its general applicability.

Takeaways, Limitations

Takeaways:
Proposal of Scalable Option Learning (SOL) algorithm.
Hierarchical agent training using 30 billion frames of experience in NetHack, outperforming existing methods.
Also verified in MiniHack and Mujoco environments.
Achieves ~35x higher throughput compared to existing hierarchical methods.
Limitations:
No specific Limitations mentioned in the paper.
👍