QC_SANE: Robust Control in DRL Using Quantile Critic With Spiking Actor and Normalized Ensemble.
IEEE Trans Neural Netw Learn Syst
; 34(9): 6656-6662, 2023 Sep.
Article
em En
| MEDLINE
| ID: mdl-34874871
Recently introduced deep reinforcement learning (DRL) techniques in discrete-time have resulted in significant advances in online games, robotics, and so on. Inspired from recent developments, we have proposed an approach referred to as Quantile Critic with Spiking Actor and Normalized Ensemble (QC_SANE) for continuous control problems, which uses quantile loss to train critic and a spiking neural network (NN) to train an ensemble of actors. The NN does an internal normalization using a scaled exponential linear unit (SELU) activation function and ensures robustness. The empirical study on multijoint dynamics with contact (MuJoCo)-based environments shows improved training and test results than the state-of-the-art approach: population coded spiking actor network (PopSAN).
Texto completo:
1
Base de dados:
MEDLINE
Idioma:
En
Revista:
IEEE Trans Neural Netw Learn Syst
Ano de publicação:
2023
Tipo de documento:
Article