Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Intention-Conditioned Flow Occupancy Models

Created by
  • Haebom

Author

Chongyi Zheng, Seohong Park, Sergey Levine, Benjamin Eysenbach

Outline

This paper presents a method for leveraging large-scale pre-trained models in the field of reinforcement learning (RL). Specifically, we propose InFOM (intention-conditioned flow occupancy models), probabilistic models that leverage flow matching to predict future states in RL environments where temporal dependencies are crucial. InFOM enhances the model's expressiveness by incorporating latent variables that capture user intent, enabling generalized policy improvement. We demonstrate that InFOM outperforms other pre-training methods on 36 state-based and 4 image-based benchmark tasks.

Takeaways, Limitations

Takeaways:
It contributes to improving sample efficiency and robustness, which are key challenges in the field of RL.
We apply the approach of pre-training large-scale models to adapt and fine-tune them for specific tasks to RL.
Model complex future state distributions using flow matching.
Enhance the expressive power of your model with latent variables that capture user intent.
It shows higher performance than existing methods in various benchmark environments.
Limitations:
There is no specific mention of Limitations in the paper itself.
👍