强化学习
计算机科学
理论(学习稳定性)
噪音(视频)
正规化(语言学)
降噪
还原(数学)
人工智能
数学优化
机器学习
数学
几何学
图像(数学)
作者
Shuai Han,Wenbo Zhou,Jing Liu,Shuai Lü
标识
DOI:10.1016/j.eswa.2022.117343
摘要
Deep reinforcement learning has been applied more and more widely nowadays, especially in various complex control tasks. Noise is vital to reinforcement learning exploration. The appropriate noise not only avoids local optimum solutions due to the lack of exploration, but also prevents unstable learning caused by excessive perturbations. Noisy networks can bring more efficient exploration to reinforcement learning. They enable agents to take actions more randomly at the start of training and tend to make the agents produce stable outputs in the subsequent learning. However, this tendency is not always enough to find a stable policy for an agent, which decreases the efficiency and stability. To address this issue, this paper proposes NROWAN-DQN, i.e., Noise Reduction and Online Weight Adjustment NoisyNet-DQN. Firstly, we develop a novel noise regularization method for NoisyNet-DQN to reduce the noise of output layers. Secondly, we design an online weight adjustment strategy for noise reduction. Finally, we evaluate this algorithm in four standard domains and analyze properties of hyper-parameters. Our results show that NROWAN-DQN outperforms prior algorithms in all these domains. In addition, NROWAN-DQN also shows better stability. The variance of the NROWAN-DQN score is significantly reduced, especially in some action-sensitive environments. This means that in some environments where high stability is required, NROWAN-DQN will be more appropriate than NoisyNet-DQN.
科研通智能强力驱动
Strongly Powered by AbleSci AI