强化学习
学习迁移
下部结构
计算机科学
公制(单位)
人工智能
钢筋
机器学习
材料科学
工程类
结构工程
复合材料
运营管理
作者
Peihua Chai,Bilian Chen,Yifeng Zeng,Shenbao Yu
出处
期刊:Neurocomputing
[Elsevier BV]
日期:2024-06-15
卷期号:598: 128071-128071
标识
DOI:10.1016/j.neucom.2024.128071
摘要
Transfer reinforcement learning has gained significant traction in recent years as a critical research area, focusing on bolstering agents' decision-making prowess by harnessing insights from analogous tasks. The primary transfer learning method involves identifying the appropriate source domains, sharing specific knowledge structures and subsequently transferring the shared knowledge to novel tasks. However, existing transfer methods exhibit a pronounced dependency on high task similarity and an abundance of source data. Consequently, we attempt to formulate a more efficacious approach that optimally exploits the previous learning experiences to direct an agent's exploration as it learns new tasks. Specifically, we introduce a novel transfer learning paradigm rooted within the distance measure in the Markov chain, denoted as Distance Measure Substructure Transfer Reinforcement Learning (DMS-TRL). The core idea involves partitioning the Markov chain into the most basic small Markov units, which contain basic information about the agent's transfer between two states, and then followed by employing a new distance measure technique to find the most similar structure, which is also the most suitable for transfer. Finally, we propose a policy transfer method to transfer knowledge through the Q table from the selected Markov unit to the target task. Through a series of experiments conducted on discrete Gridworld scenarios, we compare our approach with state-of-the-art learning methods. The results clearly illustrate that DMS-TRL can adeptly identify optimal policy in target tasks, exhibiting swifter convergence.
科研通智能强力驱动
Strongly Powered by AbleSci AI