Your browser doesn't support javascript.
loading
Optimal Tracking Control of a Nonlinear Multiagent System Using Q-Learning via Event-Triggered Reinforcement Learning.
Wang, Ziwei; Wang, Xin; Tang, Yijie; Liu, Ying; Hu, Jun.
Afiliação
  • Wang Z; College of Electronic and Information Engineering, Southwest University, Chongqing 400700, China.
  • Wang X; College of Electronic and Information Engineering, Southwest University, Chongqing 400700, China.
  • Tang Y; College of Electronic and Information Engineering, Southwest University, Chongqing 400700, China.
  • Liu Y; College of Electronic and Information Engineering, Southwest University, Chongqing 400700, China.
  • Hu J; College of Electronic and Information Engineering, Southwest University, Chongqing 400700, China.
Entropy (Basel) ; 25(2)2023 Feb 05.
Article em En | MEDLINE | ID: mdl-36832665
ABSTRACT
This article offers an optimal control tracking method using an event-triggered technique and the internal reinforcement Q-learning (IrQL) algorithm to address the tracking control issue of unknown nonlinear systems with multiple agents (MASs). Relying on the internal reinforcement reward (IRR) formula, a Q-learning function is calculated, and then the iteration IRQL method is developed. In contrast to mechanisms triggered by time, an event-triggered algorithm reduces the rate of transmission and computational load, since the controller may only be upgraded when the predetermined triggering circumstances are met. In addition, in order to implement the suggested system, a neutral reinforce-critic-actor (RCA) network structure is created that may assess the indices of performance and online learning of the event-triggering mechanism. This strategy is intended to be data-driven without having in-depth knowledge of system dynamics. We must develop the event-triggered weight tuning rule, which only modifies the parameters of the actor neutral network (ANN) in response to triggering cases. In addition, a Lyapunov-based convergence study of the reinforce-critic-actor neutral network (NN) is presented. Lastly, an example demonstrates the accessibility and efficiency of the suggested approach.
Palavras-chave

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Idioma: En Revista: Entropy (Basel) Ano de publicação: 2023 Tipo de documento: Article País de afiliação: China

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Idioma: En Revista: Entropy (Basel) Ano de publicação: 2023 Tipo de documento: Article País de afiliação: China