Abstract
Multi-robot learning is a challenging task not only because of large and continuous state/action spaces, but also uncertainty and partial observability during learning. This paper presents a distributed policy gradient reinforcement learning (PGRL) methodology of a multi-robot system using neural network as the function approximator. This distributed PGRL algorithm enables each robot to independently decide its policy, which is, however, affected by all the other robots. Neural network is used to generalize over continuous state space as well as discrete/continuous action spaces. A case study on leader-follower formation application is performed to demonstrate the effectiveness of the proposed learning method. © 2008 IEEE.
| Original language | English |
|---|---|
| Title of host publication | Proceedings of the 2008 IEEE International Conference on Information and Automation, ICIA 2008 |
| Pages | 113-118 |
| DOIs | |
| Publication status | Published - 2008 |
| Event | 2008 IEEE International Conference on Information and Automation, ICIA 2008 - Zhangjiajie, Hunan, China Duration: 20 Jun 2008 → 23 Jun 2008 |
Conference
| Conference | 2008 IEEE International Conference on Information and Automation, ICIA 2008 |
|---|---|
| Place | China |
| City | Zhangjiajie, Hunan |
| Period | 20/06/08 → 23/06/08 |