Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Learning Adaptive Dexterous Grasping from Single Demonstrations

Created by
  • Haebom

Author

Liangzhi Shi, Yulin Liu, Lingqi Zeng, Bo Ai, Zhengdong Hong, Hao Su

Outline

AdaDexGrasp is a framework that efficiently learns skilled grasping techniques from limited human demonstrations and adaptively applies them based on user instructions. It learns multiple grasping techniques from a single human demonstration and selects the most appropriate technique using a vision-language model (VLM). To increase sample efficiency, it proposes a trajectory-following reward that guides reinforcement learning (RL) toward a state closer to human demonstrations. It also learns beyond a single demonstration through curriculum learning, which incrementally increases the number of object pose variations. Upon deployment, the VLM searches for appropriate techniques based on user instructions, connecting low-level learning techniques with high-level intent. Evaluations in simulations and real-world environments demonstrate that it significantly improves RL efficiency and enables the learning of human-like grasping strategies across a variety of object configurations. Zero-shot transfer of the learned policy to the real PSYONIC Ability Hand achieves a 90% success rate on objects, significantly outperforming the baseline.

Takeaways, Limitations

Takeaways:
Efficient and skilled phage skills can be learned from limited human demonstrations.
Reflecting high-level intent through context-based technology selection using vision-language models.
Robust skill learning and generalization through curriculum learning.
Successful zero-shot transfer to actual robot hand.
Learning human-like phage strategies.
Limitations:
Lack of detailed description of the specific structure and performance of the proposed VLM.
Limitations in generalization performance across a variety of objects and situations.
Additional testing and validation in real-world environments is required.
Lack of discussion on setting optimal parameters for trajectory tracking compensation.
👍