Coordinated Multi-Agent Patrolling with State-Dependent Cost Rates : Asymptotically Optimal Policies for Large-Scale Systems

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

1 Scopus Citations
View graph of relations

Author(s)

Related Research Unit(s)

Detail(s)

Original languageEnglish
Journal / PublicationIEEE Transactions on Automatic Control
Publication statusOnline published - 13 Dec 2024

Abstract

We study a large-scale patrol problem with state-dependent costs and multi-agent coordination. We consider heterogeneous agents, rather general reward functions, and the capabilities of tracking agents' trajectories. We model the problem as a discrete-time Markov decision process consisting of parallel stochastic processes. The problem exhibits an excessively large state space, which increases exponentially in the number of agents and the size of patrol region. By randomizing all the action variables, we relax and decompose the problem into multiple sub-problems, each of which can be solved independently and lead to scalable heuristics applicable to the original problem. Unlike the past studies assuming relatively simple structures of the underlying stochastic process, here, tracking the patrol trajectories involves stronger dependencies between the stochastic processes, leading to entirely different state and action spaces and transition kernels, rendering the existing methods inapplicable or impractical. Furthermore, we prove that the performance deviation between the proposed policies and the possible optimal solution diminishes exponentially in the problem size. © 2024 IEEE.

Research Area(s)

  • Asymptotic optimality, multi-agent patrolling, restless bandit