Abstract
Achieving a connected, collision-free and time-efficient coverage in unknown environments is challenging for multi-agent systems. Particularly, agents with second-order dynamics are supposed to efficiently search and reach the optimal deployment positions over targets whose distribution is unknown, while preserving the distributed connectivity and avoiding collision. In this paper, a safe reinforcement learning based shield method is proposed for unknown environment exploration while correcting actions of agents for safety guarantee and avoiding invalid samples into policy updating. The shield is achieved distributively by a control barrier function and its validity is proved in theory. Moreover, policies of the optimal coverage are centrally learned via reward engineering and executed distributively. Numerical results show that the proposed approach not only achieves zero safety violations during training, but also speeds up the convergence of learning.
Original language | English |
---|---|
Title of host publication | Proceedings of the 62nd IEEE Conference on Decision and Control (CDC 2023), Singapore, 13-15 December 2023 |
Place of Publication | U.S. |
Publisher | IEEE |
Pages | 3415-3420 |
Number of pages | 6 |
ISBN (Electronic) | 9798350301243 |
DOIs | |
Publication status | Published - 2023 |
Event | IEEE Conference on Decision & Control - Marina Bay Sands, Singapore, Singapore Duration: 13 Dec 2023 → 15 Dec 2023 Conference number: 62nd |
Conference
Conference | IEEE Conference on Decision & Control |
---|---|
Country/Territory | Singapore |
City | Singapore |
Period | 13/12/23 → 15/12/23 |