Abstract
This paper proposes a novel reinforcement learning approach to solve the optimal output synchronization problem for discrete-time heterogeneous multi-agent systems. Different from existing learning methods, the optimal control protocol is obtained to guarantee zero synchronization error by solving the augmented algebraic Riccati equations (AREs). The proposed adaptive dynamic programming (ADP) method can stabilize the output synchronization error and solve the output regulator equations implicitly. To eliminate the dependency on information of system dynamics, an online Q-function-based policy iteration (PI) algorithm is developed. Finally, a numerical example is provided to demonstrate the advantages of the proposed ADP over traditional ADP in terms of synchronization performance.
| Original language | English |
|---|---|
| Title of host publication | 2024 American Control Conference, ACC 2024 |
| Publisher | IEEE |
| Pages | 3656-3661 |
| Number of pages | 6 |
| ISBN (Electronic) | 9798350382655 |
| ISBN (Print) | 9798350382662 |
| DOIs | |
| Publication status | Published - 2024 |
| Externally published | Yes |
| Event | 2024 American Control Conference, ACC 2024 - Toronto, Canada Duration: 10 Jul 2024 → 12 Jul 2024 |
Publication series
| Name | Proceedings of the American Control Conference |
|---|---|
| ISSN (Print) | 0743-1619 |
Conference
| Conference | 2024 American Control Conference, ACC 2024 |
|---|---|
| Country/Territory | Canada |
| City | Toronto |
| Period | 10/07/24 → 12/07/24 |
Bibliographical note
Publisher Copyright:© 2024 AACC.
Funding
This work was supported by the Guangdong Basic and Applied Basic Research Foundation under project 2023A1515011981, the Shenzhen Science and Technology Program under projects JCYJ20220818102416036 and partly supported by the Natural Sciences and Engineering Research Council of Canada (NSERC).