Deep Reinforcement Learning for Image-Based Multi-Agent Coverage Path Planning
Research output: Chapters, Conference Papers, Creative and Literary Works › RGC 32 - Refereed conference paper (with host publication) › peer-review
Author(s)
Related Research Unit(s)
Detail(s)
Original language | English |
---|---|
Title of host publication | 2023 IEEE 98th Vehicular Technology Conference (VTC2023-Fall) - Proceedings |
Publisher | Institute of Electrical and Electronics Engineers, Inc. |
Number of pages | 7 |
ISBN (electronic) | 979-8-3503-2928-5 |
ISBN (print) | 979-8-3503-2929-2 |
Publication status | Published - 2023 |
Publication series
Name | IEEE Vehicular Technology Conference |
---|---|
ISSN (Print) | 1090-3038 |
ISSN (electronic) | 2577-2465 |
Conference
Title | 98th IEEE Vehicular Technology Conference (VTC 2023-Fall) |
---|---|
Place | Hong Kong |
Period | 10 - 13 October 2023 |
Link(s)
Abstract
Image-based Multi-Agent Coverage Path Planning (MACPP) utilizes images as input to control multiple agents touring all nodes in a map, minimizing task duration and node revisiting. State-of-the-art (SOTA) studies have applied Multi-Agent Deep Reinforcement Learning (MADRL) to automate MACPP, primarily focusing on minimizing task duration. However, these approaches overlook the issue of repeated node visits, resulting in longer task durations and limited real-world applicability. To tackle this challenge, we develop a novel MADRL solution, referred to as MADRL with Mask Soft Attention, to minimize task duration and node re-visiting simultaneously. Our method uses mask soft attention to extract key features from raw image observations while masking task-independent features, reducing computational complexity and improving sample efficiency. We also cascade a multi-actor-critic architecture to accommodate even more agents with ease. Each agent is equipped with an actor to learn an action policy, and a shared critic evaluates a state value. To validate our approach, we implement seven SOTA MADRL methods in the MACPP area as baselines. Simulation results show that our method significantly outperforms the baselines regarding task duration and the number of times the node is repeatedly visited. © 2023 IEEE.
Research Area(s)
- deep reinforcement learning, multi-agent coverage path planning, multi-actor-critic, mask soft attention
Citation Format(s)
Deep Reinforcement Learning for Image-Based Multi-Agent Coverage Path Planning. / Xu, Meng; She, Yechao; Jin, Yang et al.
2023 IEEE 98th Vehicular Technology Conference (VTC2023-Fall) - Proceedings. Institute of Electrical and Electronics Engineers, Inc., 2023. (IEEE Vehicular Technology Conference).
2023 IEEE 98th Vehicular Technology Conference (VTC2023-Fall) - Proceedings. Institute of Electrical and Electronics Engineers, Inc., 2023. (IEEE Vehicular Technology Conference).
Research output: Chapters, Conference Papers, Creative and Literary Works › RGC 32 - Refereed conference paper (with host publication) › peer-review