Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Human2LocoMan: Learning Versatile Quadrupedal Manipulation with Human Pretraining

Created by
  • Haebom

Author

Yaru Niu, Yunzhe Zhang, Mingyang Yu, Changyi Lin, Chenhao Li, Yikai Wang, Yuxiang Yang, Wenhao Yu, Tingnan Zhang, Zhenzhen Li, Jonathan Francis, Bingqing Chen, Jie Tan, Ding Zhao

Outline

This paper presents a solution to the problem of equipping a quadruped robot with scalable autonomous and multi-purpose manipulation skills that demonstrate impressive locomotion capabilities in diverse environments. We introduce a cross-implementation imitation learning system that leverages data collected from a human and a quadruped robot with multiple manipulation modes, LocoMan. We develop a teleoperation and data acquisition pipeline that integrates and modularizes the observation and action spaces of humans and robots, and propose an efficient modular architecture that supports joint learning and pre-training using structured mode-aligned data across multiple implementations. We also construct the first manipulation dataset for LocoMan robots that covers a variety of household tasks in both single- and dual-handed modes, complementing the corresponding human dataset. Our validation results on six real-world manipulation tasks show an overall improvement of 41.9% over the baseline, and an average success rate of 79.7% in the out-of-distribution setting. Pre-training with human data contributes an overall improvement of 38.6% and an 82.7% success rate improvement in the out-of-distribution setting, consistently achieving better performance with only half the robot data. The code, hardware, and data have been made public ( https://human2bots.github.io ).

Takeaways, Limitations

Takeaways:
We significantly improved the manipulation performance of quadruped robots through a cross-implementation imitation learning system that integrates and leverages human and robot data.
We improved data efficiency through efficient modular architecture and pre-training strategy.
We show that pre-training on human data can improve performance while reducing the amount of robot data.
We present successful validation results for real-world tasks.
We made our research reproducible and scalable by making our code, hardware, and data open.
Limitations:
Since we currently use a dataset limited to home tasks, generalization performance evaluation for various environments and tasks is required.
Performance may be affected by the quality of the human demo.
Further research is needed into robots' ability to cope with unpredictable situations and obstacles.
Further research is needed to learn more complex and diverse manipulation skills.
👍