可解释性
过度拟合
人工智能
强化学习
计算机科学
步伐
多样性(控制论)
机器学习
深度学习
人工神经网络
钥匙(锁)
深层神经网络
网络体系结构
价值网络
价值(数学)
业务
营销
地理
计算机安全
大地测量学
商业模式
作者
Raghuram Mandyam Annasamy,Katia Sycara
出处
期刊:Proceedings of the ... AAAI Conference on Artificial Intelligence
[Association for the Advancement of Artificial Intelligence (AAAI)]
日期:2019-07-17
卷期号:33 (01): 4561-4569
被引量:51
标识
DOI:10.1609/aaai.v33i01.33014561
摘要
Deep reinforcement learning techniques have demonstrated superior performance in a wide variety of environments. As improvements in training algorithms continue at a brisk pace, theoretical or empirical studies on understanding what these networks seem to learn, are far behind. In this paper we propose an interpretable neural network architecture for Q-learning which provides a global explanation of the model’s behavior using key-value memories, attention and reconstructible embeddings. With a directed exploration strategy, our model can reach training rewards comparable to the state-of-the-art deep Q-learning models. However, results suggest that the features extracted by the neural network are extremely shallow and subsequent testing using out-of-sample examples shows that the agent can easily overfit to trajectories seen during training.
科研通智能强力驱动
Strongly Powered by AbleSci AI