Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Decentralized Aerial Manipulation of a Cable-Suspended Load using Multi-Agent Reinforcement Learning

Created by
  • Haebom

Author

Jack Zeng, Andreu Matoses Gimenez, Eugene Vinitsky, Javier Alonso-Mora, Sihao Sun

Outline

This paper presents the first distributed control approach for manipulating a cable-suspended object in a real-world environment with six degrees of freedom (DOF) using multiple small airborne vehicles (MAVs). Multi-agent reinforcement learning (MARL) is used to train a high-level control policy for each MAV. Unlike existing centralized control approaches, the proposed approach does not require global state information, inter-MAV communication, or information about neighboring MAVs. Instead, the agents implicitly communicate only with the object's attitude information, providing high scalability and flexibility. Furthermore, it significantly reduces computational overhead during inference, enabling onboard deployment. A novel motion space design using linear acceleration and attitude rate, combined with a robust sub-controller, enables reliable simulation-to-real-world transfer despite significant uncertainty due to cable tension during dynamic 3D motion. This is validated through various real-world experiments, including full attitude control under load model uncertainty, demonstrating setpoint tracking performance comparable to state-of-the-art centralized methods. Furthermore, the proposed approach demonstrates cooperation between agents with heterogeneous control policies and robustness against the complete loss of a single MAV during flight.

Takeaways, Limitations

Takeaways:
The first distributed control scheme for six-degree-of-freedom manipulation of cable-suspended objects is presented.
High scalability and flexibility without global state information or inter-MAV communication.
Reduced inference time computational costs to enable onboard deployment.
Validation of comparable performance to state-of-the-art centralized methods through experiments in real-world environments.
Robustness against heterogeneous policies and MAV loss.
Limitations:
There is a lack of specific mention of the general Limitations or vulnerability of the method presented in this paper to specific situations.
Further research is needed to explore its applicability to diverse environments or more complex tasks.
Due to the lack of detailed information on the experimental environment, reproducibility needs to be reviewed.
👍