Abstract
This article investigates the optimal control problem with disturbance rejection for discrete-time multi-agent systems under cooperative and non-cooperative graphical games frameworks. Given the practical challenges of obtaining accurate models, Q-function-based policy iteration methods are proposed to seek the Nash equilibrium solution for the cooperative graphical game and the distributed minmax solution for the non-cooperative graphical game. To implement these methods online, two reinforcement learning frameworks are developed, an actor-disturber-critic structure for the cooperative graphical game and an actor-adversary-disturber-critic structure for the non-cooperative graphical game. The stability of the proposed methods is rigorously analyzed, and simulation results are provided to illustrate the effectiveness of the proposed methods.
| Original language | English |
|---|---|
| Pages (from-to) | 585-601 |
| Number of pages | 17 |
| Journal | International Journal of Robust and Nonlinear Control |
| Volume | 36 |
| Issue number | 2 |
| Early online date | 20 Aug 2025 |
| DOIs | |
| Publication status | Published - 25 Jan 2026 |
| Externally published | Yes |
Bibliographical note
Publisher Copyright:© 2025 John Wiley & Sons Ltd.
Funding
This work was supported by the National Key R&D Program of China under Grant 2022YFB4700200, the Guangdong Basic and Applied Basic Research Foundation under project 2023A1515011981, the Shenzhen Science and Technology Program under projects JCYJ20220818102416036 and RCJC20210609104400005, and partly by NSERC.
Keywords
- disturbance rejection
- multi-agent system
- Nash equilibrium
- reinforcement learning