1
|
Xue W, Zhan S, Wu Z, Chen Y, Huang J. Distributed multi-agent collision avoidance using robust differential game. ISA TRANSACTIONS 2023; 134:95-107. [PMID: 36182609 DOI: 10.1016/j.isatra.2022.09.012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/05/2022] [Revised: 09/06/2022] [Accepted: 09/06/2022] [Indexed: 06/16/2023]
Abstract
This paper proposes a novel robust differential game scheme to solve the collision avoidance problem for networked multi-agent systems (MASs), subject to linear dynamics, external disturbances and limited observation capabilities. Compared with the existing differential game approaches only considering obstacle avoidance objectives, we explicitly incorporate the trajectory optimization target by penalizing the deviation from reference trajectories, based on the artificial potential field (APF) concept. It is proved that the strategies of each agent defined by individual optimization problems will converge to a local robust Nash equilibrium (R-NE), which further, with a fixed strong connection topology, will converge to the global R-NE. Additionally, to cope with the limited observation for MASs, local robust feedback control strategies are constructed based on the best approximate cost function and distributed robust Hamilton-Jacobi-Isaacs (DR-HJI) equations, which does not require global information of agents as in the traditional Riccati equation form. The feedback gains of the control strategies are found via the ant colony optimization (ACO) algorithm with a non-dominant sorting structure with convergence guarantees. Finally, simulation results are provided to verify the efficacy and robustness of the novel scheme. The agents arrived at the targeted position collision-free with a reduced arrival time, and reached the targeted positions under disturbance.
Collapse
Affiliation(s)
- Wenyan Xue
- The College of Electrical Engineering and Automation, Fuzhou University, Fuzhou, 350108, China; The Institute of 5G+ Industrial Internet, Fuzhou University, Fuzhou, 350108, China
| | - Siyuan Zhan
- Department of Electronic Engineering, Maynooth University, Maynooth, W23 F2K8, Ireland
| | - Zhihong Wu
- The College of Electrical Engineering and Automation, Fuzhou University, Fuzhou, 350108, China; The Institute of 5G+ Industrial Internet, Fuzhou University, Fuzhou, 350108, China
| | - Yutao Chen
- The College of Electrical Engineering and Automation, Fuzhou University, Fuzhou, 350108, China; The Institute of 5G+ Industrial Internet, Fuzhou University, Fuzhou, 350108, China
| | - Jie Huang
- The College of Electrical Engineering and Automation, Fuzhou University, Fuzhou, 350108, China; The Institute of 5G+ Industrial Internet, Fuzhou University, Fuzhou, 350108, China
| |
Collapse
|
2
|
Wu Q, Wu Y, Wang Y. Integral Reinforcement-Learning-Based Optimal Containment Control for Partially Unknown Nonlinear Multiagent Systems. ENTROPY (BASEL, SWITZERLAND) 2023; 25:221. [PMID: 36832588 PMCID: PMC9955993 DOI: 10.3390/e25020221] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/20/2022] [Revised: 01/21/2023] [Accepted: 01/22/2023] [Indexed: 06/18/2023]
Abstract
This paper focuses on the optimal containment control problem for the nonlinear multiagent systems with partially unknown dynamics via an integral reinforcement learning algorithm. By employing integral reinforcement learning, the requirement of the drift dynamics is relaxed. The integral reinforcement learning method is proved to be equivalent to the model-based policy iteration, which guarantees the convergence of the proposed control algorithm. For each follower, the Hamilton-Jacobi-Bellman equation is solved by a single critic neural network with a modified updating law which guarantees the weight error dynamic to be asymptotically stable. Through using input-output data, the approximate optimal containment control protocol of each follower is obtained by applying the critic neural network. The closed-loop containment error system is guaranteed to be stable under the proposed optimal containment control scheme. Simulation results demonstrate the effectiveness of the presented control scheme.
Collapse
Affiliation(s)
| | | | - Yonghua Wang
- School of Automation, Guangdong University of Technology, Guangzhou 510006, China
| |
Collapse
|