Safe Multi-Agent Reinforcement Learning for Formation Control Without Individual Reference Targets




Authors:

M. Dawood, S. Pan, N. Dengler, S. Zhou, A. Schoellig, M. Bennewitz

Type:

Preprint

Published in:

ArXiv preprint

Year:

2023

Links:

PreprintVideo

BibTex String

@article{dawood2023safe,
title={Safe Multi-Agent Reinforcement Learning for Formation Control without Individual Reference Targets},
author={Dawood, Murad and Pan, Sicong and Dengler, Nils and Zhou, Siqi and Schoellig, Angela P and Bennewitz, Maren},
journal={arXiv preprint arXiv:2312.12861},
year={2023}
}
Topic

Abstract:

In recent years, formation control of unmanned vehicles has received considerable interest, driven by the progress in autonomous systems and the imperative for multiple vehicles to carry out diverse missions. In this paper, we address the problem of behavior-based formation control of mobile robots, where we use safe multi-agent reinforcement learning~(MARL) to ensure the safety of the robots by eliminating all collisions during training and execution. To ensure safety, we implemented distributed model predictive control safety filters to override unsafe actions. We focus on achieving behavior-based formation without having individual reference targets for the robots, and instead use targets for the centroid of the formation. This formulation facilitates the deployment of formation control on real robots and improves the scalability of our approach to more robots. The task cannot be addressed through optimization-based controllers without specific individual reference targets for the robots and information about the relative locations of each robot to the others. That is why, for our formulation we use MARL to train the robots. Moreover, in order to account for the interactions between the agents, we use attention-based critics to improve the training process. We train the agents in simulation and later on demonstrate the resulting behavior of our approach on real Turtlebot robots. We show that despite the agents having very limited information, we can still safely achieve the desired behavior.