Optimal model-free output synchronization of heterogeneous systems using off-policy reinforcement learning

Journal Article (2016)
Authors

H Modares (University of Texas at Arlington Research Inst., Fort Worth, TX, USA)

S.P. Nageshrao (OLD Intelligent Control & Robotics)

G.A. Lopes (OLD Intelligent Control & Robotics)

R Babuška (OLD Intelligent Control & Robotics)

FL Lewis (Northeastern University China, University of Texas at Arlington Research Inst., Fort Worth, TX, USA)

Research Group
OLD Intelligent Control & Robotics
More Info
expand_more
Publication Year
2016
Language
English
Research Group
OLD Intelligent Control & Robotics
Volume number
71
Pages (from-to)
334-341
DOI:
https://doi.org/10.1016/j.automatica.2016.05.017

Abstract

This paper considers optimal output synchronization of heterogeneous linear multi-agent systems. Standard approaches to output synchronization of heterogeneous systems require either the solution of the output regulator equations or the incorporation of a p-copy of the leader’s dynamics in the controller of each agent. By contrast, in this paper neither one is needed. Moreover, here both the leader’s and the follower’s dynamics are assumed to be unknown. First, a distributed adaptive observer is designed to estimate the leader’s state for each agent. The output synchronization problem is then formulated as an optimal control problem and a novel model-free off-policy reinforcement learning algorithm is developed to solve the optimal output synchronization problem online in real time. It is shown that this optimal distributed approach implicitly solves the output regulation equations without actually doing so.
Simulation results are provided to verify the effectiveness of the proposed approach.

No files available

Metadata only record. There are no files for this record.