Constrained coverage of unknown environment using safe reinforcement learning

Yunlin Zhang, Junjie You, Lei Shi, Jinliang Shao, Wei Xing Zheng

Research output: Chapter in Book / Conference PaperConference Paperpeer-review

1 Citation (Scopus)

Abstract

Achieving a connected, collision-free and time-efficient coverage in unknown environments is challenging for multi-agent systems. Particularly, agents with second-order dynamics are supposed to efficiently search and reach the optimal deployment positions over targets whose distribution is unknown, while preserving the distributed connectivity and avoiding collision. In this paper, a safe reinforcement learning based shield method is proposed for unknown environment exploration while correcting actions of agents for safety guarantee and avoiding invalid samples into policy updating. The shield is achieved distributively by a control barrier function and its validity is proved in theory. Moreover, policies of the optimal coverage are centrally learned via reward engineering and executed distributively. Numerical results show that the proposed approach not only achieves zero safety violations during training, but also speeds up the convergence of learning.
Original languageEnglish
Title of host publicationProceedings of the 62nd IEEE Conference on Decision and Control (CDC 2023), Singapore, 13-15 December 2023
Place of PublicationU.S.
PublisherIEEE
Pages3415-3420
Number of pages6
ISBN (Electronic)9798350301243
DOIs
Publication statusPublished - 2023
EventIEEE Conference on Decision & Control - Marina Bay Sands, Singapore, Singapore
Duration: 13 Dec 202315 Dec 2023
Conference number: 62nd

Conference

ConferenceIEEE Conference on Decision & Control
Country/TerritorySingapore
CitySingapore
Period13/12/2315/12/23

Fingerprint

Dive into the research topics of 'Constrained coverage of unknown environment using safe reinforcement learning'. Together they form a unique fingerprint.

Cite this