Your browser doesn't support javascript.
loading
QC_SANE: Robust Control in DRL Using Quantile Critic With Spiking Actor and Normalized Ensemble.
IEEE Trans Neural Netw Learn Syst ; 34(9): 6656-6662, 2023 Sep.
Article em En | MEDLINE | ID: mdl-34874871
Recently introduced deep reinforcement learning (DRL) techniques in discrete-time have resulted in significant advances in online games, robotics, and so on. Inspired from recent developments, we have proposed an approach referred to as Quantile Critic with Spiking Actor and Normalized Ensemble (QC_SANE) for continuous control problems, which uses quantile loss to train critic and a spiking neural network (NN) to train an ensemble of actors. The NN does an internal normalization using a scaled exponential linear unit (SELU) activation function and ensures robustness. The empirical study on multijoint dynamics with contact (MuJoCo)-based environments shows improved training and test results than the state-of-the-art approach: population coded spiking actor network (PopSAN).

Texto completo: 1 Base de dados: MEDLINE Idioma: En Revista: IEEE Trans Neural Netw Learn Syst Ano de publicação: 2023 Tipo de documento: Article

Texto completo: 1 Base de dados: MEDLINE Idioma: En Revista: IEEE Trans Neural Netw Learn Syst Ano de publicação: 2023 Tipo de documento: Article