TY - GEN
T1 - Deep Reinforcement Learning for Image-Based Multi-Agent Coverage Path Planning
AU - Xu, Meng
AU - She, Yechao
AU - Jin, Yang
AU - Wang, Jianping
PY - 2023
Y1 - 2023
N2 - Image-based Multi-Agent Coverage Path Planning (MACPP) utilizes images as input to control multiple agents touring all nodes in a map, minimizing task duration and node revisiting. State-of-the-art (SOTA) studies have applied Multi-Agent Deep Reinforcement Learning (MADRL) to automate MACPP, primarily focusing on minimizing task duration. However, these approaches overlook the issue of repeated node visits, resulting in longer task durations and limited real-world applicability. To tackle this challenge, we develop a novel MADRL solution, referred to as MADRL with Mask Soft Attention, to minimize task duration and node re-visiting simultaneously. Our method uses mask soft attention to extract key features from raw image observations while masking task-independent features, reducing computational complexity and improving sample efficiency. We also cascade a multi-actor-critic architecture to accommodate even more agents with ease. Each agent is equipped with an actor to learn an action policy, and a shared critic evaluates a state value. To validate our approach, we implement seven SOTA MADRL methods in the MACPP area as baselines. Simulation results show that our method significantly outperforms the baselines regarding task duration and the number of times the node is repeatedly visited. © 2023 IEEE.
AB - Image-based Multi-Agent Coverage Path Planning (MACPP) utilizes images as input to control multiple agents touring all nodes in a map, minimizing task duration and node revisiting. State-of-the-art (SOTA) studies have applied Multi-Agent Deep Reinforcement Learning (MADRL) to automate MACPP, primarily focusing on minimizing task duration. However, these approaches overlook the issue of repeated node visits, resulting in longer task durations and limited real-world applicability. To tackle this challenge, we develop a novel MADRL solution, referred to as MADRL with Mask Soft Attention, to minimize task duration and node re-visiting simultaneously. Our method uses mask soft attention to extract key features from raw image observations while masking task-independent features, reducing computational complexity and improving sample efficiency. We also cascade a multi-actor-critic architecture to accommodate even more agents with ease. Each agent is equipped with an actor to learn an action policy, and a shared critic evaluates a state value. To validate our approach, we implement seven SOTA MADRL methods in the MACPP area as baselines. Simulation results show that our method significantly outperforms the baselines regarding task duration and the number of times the node is repeatedly visited. © 2023 IEEE.
KW - deep reinforcement learning
KW - multi-agent coverage path planning
KW - multi-actor-critic
KW - mask soft attention
UR - http://www.scopus.com/inward/record.url?scp=85181174480&partnerID=8YFLogxK
UR - https://www.scopus.com/record/pubmetrics.uri?eid=2-s2.0-85181174480&origin=recordpage
U2 - 10.1109/VTC2023-Fall60731.2023.10333800
DO - 10.1109/VTC2023-Fall60731.2023.10333800
M3 - RGC 32 - Refereed conference paper (with host publication)
SN - 979-8-3503-2929-2
T3 - IEEE Vehicular Technology Conference
BT - 2023 IEEE 98th Vehicular Technology Conference (VTC2023-Fall) - Proceedings
PB - IEEE
T2 - 98th IEEE Vehicular Technology Conference (VTC 2023-Fall)
Y2 - 10 October 2023 through 13 October 2023
ER -