A novel policy based on action confidence limit to improve exploration efficiency in reinforcement learning

Fanghui Huang, Xinyang Deng, Yixin He, Wen Jiang

科研成果: 期刊稿件文章同行评审

13 引用 (Scopus)

摘要

Reinforcement learning has been used to solve many intelligent decision-making problems. However, reinforcement learning still faces a challenge of the low exploration efficiency problem in practice, limiting its widespread application. To address this issue, in this paper, a novel exploration policy based on Q value and exploration value is proposed. The exploration value adopts action confidence limit to measure the uncertainty of the action, which guides the agent to adaptively explore the uncertainty region of the environment. This method can improve exploration efficiency and is beneficial for the agent to make optimal decisions. Then, in order to make our proposed policy applicable to discrete and continuous environments, we combine the proposed policy with classic reinforcement learning algorithms (Q-learning and deep Q-network), and propose two novel algorithms, respectively. Moreover, the convergence of the algorithm is analyzed. Furthermore, a deep auto-encoder network model is used to establish the mapping relationship on state-action in discrete environments, which can avoid a large number of state-action pairs stored in Q-learning stage. Our proposed method can achieve adaptive and effective exploration, which is beneficial for the agent to make intelligent decisions. Finally, the results are verified in discrete and continuous simulation environments. Experimental results demonstrate that our method improves the average reward value and reduces the number of catastrophic actions.

源语言英语
文章编号119011
期刊Information Sciences
640
DOI
出版状态已出版 - 9月 2023

指纹

探究 'A novel policy based on action confidence limit to improve exploration efficiency in reinforcement learning' 的科研主题。它们共同构成独一无二的指纹。

引用此