Abstract
Topology robustness is critical to the connectivity and lifetime of large-scale Internet-of-Things (IoT) applications. To improve robustness while reducing the execution cost, the existing robustness optimization methods utilize neural learning schemes, including neural networks, deep learning, and reinforcement learning. However, insufficient exploration of reinforcement learning agents for topological environments is likely to yield local optima. Moreover, convergence speed is influenced by the sparse reward problem generated while exploring topological environments. To address these problems, this study proposes a self-adaptive robustness optimization method with an evolutionary multi-agent for IoT topology (ROMEM). ROMEM introduces a new multi-agent co-evolution scheme that leverages a non-deterministic strategy to extend the exploration in multi-directions, enabling the reinforcement learning agent to transcend local optima. Furthermore, ROMEM presents a novel distributed training mechanism for multiple agents to accelerate convergence. Experimental results demonstrate that ROMEM can achieve multi-directional collaborative training and outperform other state-of-the-art learning-based robustness optimization methods in terms of convergence efficiency and robustness. © 2023 IEEE.
| Original language | English |
|---|---|
| Pages (from-to) | 1346-1361 |
| Journal | IEEE/ACM Transactions on Networking |
| Volume | 32 |
| Issue number | 2 |
| Online published | 13 Oct 2023 |
| DOIs | |
| Publication status | Published - Apr 2024 |
Funding
This work was supported in part by the National Science Fund for Distinguished Young Scholars of China under Grant 62325208, in part by the Joint Funds of the National Natural Science Foundation of China under Grant U2001204, and in part by the National Natural Science Foundation of China under Grant 62272339.
Research Keywords
- evolutionary reinforcement learning
- Internet of Things
- network topology
- robustness optimization
- soft actor-critic