Deep Reinforcement Learning for Image-Based Multi-Agent Coverage Path Planning

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

1 Scopus Citations
View graph of relations

Related Research Unit(s)

Detail(s)

Original languageEnglish
Title of host publication2023 IEEE 98th Vehicular Technology Conference (VTC2023-Fall) - Proceedings
PublisherInstitute of Electrical and Electronics Engineers, Inc.
Number of pages7
ISBN (electronic)979-8-3503-2928-5
ISBN (print)979-8-3503-2929-2
Publication statusPublished - 2023

Publication series

NameIEEE Vehicular Technology Conference
ISSN (Print)1090-3038
ISSN (electronic)2577-2465

Conference

Title98th IEEE Vehicular Technology Conference (VTC 2023-Fall)
PlaceHong Kong
Period10 - 13 October 2023

Abstract

Image-based Multi-Agent Coverage Path Planning (MACPP) utilizes images as input to control multiple agents touring all nodes in a map, minimizing task duration and node revisiting. State-of-the-art (SOTA) studies have applied Multi-Agent Deep Reinforcement Learning (MADRL) to automate MACPP, primarily focusing on minimizing task duration. However, these approaches overlook the issue of repeated node visits, resulting in longer task durations and limited real-world applicability. To tackle this challenge, we develop a novel MADRL solution, referred to as MADRL with Mask Soft Attention, to minimize task duration and node re-visiting simultaneously. Our method uses mask soft attention to extract key features from raw image observations while masking task-independent features, reducing computational complexity and improving sample efficiency. We also cascade a multi-actor-critic architecture to accommodate even more agents with ease. Each agent is equipped with an actor to learn an action policy, and a shared critic evaluates a state value. To validate our approach, we implement seven SOTA MADRL methods in the MACPP area as baselines. Simulation results show that our method significantly outperforms the baselines regarding task duration and the number of times the node is repeatedly visited. © 2023 IEEE.

Research Area(s)

  • deep reinforcement learning, multi-agent coverage path planning, multi-actor-critic, mask soft attention

Citation Format(s)

Deep Reinforcement Learning for Image-Based Multi-Agent Coverage Path Planning. / Xu, Meng; She, Yechao; Jin, Yang et al.
2023 IEEE 98th Vehicular Technology Conference (VTC2023-Fall) - Proceedings. Institute of Electrical and Electronics Engineers, Inc., 2023. (IEEE Vehicular Technology Conference).

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review