Distributed neural network-based policy gradient reinforcement learning for multi-robot formations

Wen Shang, Dong Sun

    Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

    3 Citations (Scopus)

    Abstract

    Multi-robot learning is a challenging task not only because of large and continuous state/action spaces, but also uncertainty and partial observability during learning. This paper presents a distributed policy gradient reinforcement learning (PGRL) methodology of a multi-robot system using neural network as the function approximator. This distributed PGRL algorithm enables each robot to independently decide its policy, which is, however, affected by all the other robots. Neural network is used to generalize over continuous state space as well as discrete/continuous action spaces. A case study on leader-follower formation application is performed to demonstrate the effectiveness of the proposed learning method. © 2008 IEEE.
    Original languageEnglish
    Title of host publicationProceedings of the 2008 IEEE International Conference on Information and Automation, ICIA 2008
    Pages113-118
    DOIs
    Publication statusPublished - 2008
    Event2008 IEEE International Conference on Information and Automation, ICIA 2008 - Zhangjiajie, Hunan, China
    Duration: 20 Jun 200823 Jun 2008

    Conference

    Conference2008 IEEE International Conference on Information and Automation, ICIA 2008
    PlaceChina
    CityZhangjiajie, Hunan
    Period20/06/0823/06/08

    Fingerprint

    Dive into the research topics of 'Distributed neural network-based policy gradient reinforcement learning for multi-robot formations'. Together they form a unique fingerprint.

    Cite this