强化学习
趋同(经济学)
数学优化
差异(会计)
计算机科学
高斯分布
文件夹
动态规划
最大熵原理
熵(时间箭头)
数学
人工智能
经济
财务
会计
物理
量子力学
经济增长
作者
Min Dai,Yuchao Dong,Yanwei Jia
摘要
Abstract We study a dynamic mean‐variance portfolio optimization problem under the reinforcement learning framework, where an entropy regularizer is introduced to induce exploration. Due to the time–inconsistency involved in a mean‐variance criterion, we aim to learn an equilibrium policy. Under an incomplete market setting, we obtain a semi‐analytical, exploratory, equilibrium mean‐variance policy that turns out to follow a Gaussian distribution. We then focus on a Gaussian mean return model and propose a reinforcement learning algorithm to find the equilibrium policy. Thanks to a thoroughly designed policy iteration procedure in our algorithm, we prove the convergence of our algorithm under mild conditions, despite that dynamic programming principle and the usual policy improvement theorem failing to hold for an equilibrium policy. Numerical experiments are given to demonstrate our algorithm. The design and implementation of our reinforcement learning algorithm apply to a general market setup.
科研通智能强力驱动
Strongly Powered by AbleSci AI