AVERAGED SOFT ACTOR-CRITIC FOR DEEP REINFORCEMENT LEARNING

Averaged Soft Actor-Critic for Deep Reinforcement Learning

Averaged Soft Actor-Critic for Deep Reinforcement Learning

Blog Article

With the advent of the era of artificial intelligence, deep reinforcement learning (DRL) has achieved unprecedented success in high-dimensional and large-scale artificial intelligence tasks.However, the insecurity and instability of the DRL algorithm have an important impact on its performance.The Soft Actor-Critic (SAC) algorithm uses advanced functions to update the policy and value network to j&d manufacturing website alleviate some of these problems.

However, SAC still has some problems.In order to reduce the error caused by the overestimation of SAC, we propose a new SAC algorithm called Averaged-SAC.By averaging the previously learned freetress deep twist crochet hair 14 inch action-state estimates, it reduces the overestimation problem of soft Q-learning, thereby contributing to a more stable training process and improving performance.

We evaluate the performance of Averaged-SAC through some games in the MuJoCo environment.The experimental results show that the Averaged-SAC algorithm effectively improves the performance of the SAC algorithm and the stability of the training process.

Report this page