强化学习
数学优化
下行风险
计算机科学
稳健优化
投资组合优化
文件夹
套利
最优化问题
人工智能
稳健性(进化)
数学
经济
财务
化学
基因
生物化学
作者
Sebastian Jaimungal,Silvana M. Pesenti,Ye Sheng Wang,Hariom Tatsat
摘要
We present a reinforcement learning (RL) approach for robust optimization of risk-aware performance criteria. To allow agents to express a wide variety of risk-reward profiles, we assess the value of a policy using rank dependent expected utility (RDEU). RDEU allows agents to seek gains, while simultaneously protecting themselves against downside risk. To robustify optimal policies against model uncertainty, we assess a policy not by its distribution but rather by the worst possible distribution that lies within a Wasserstein ball around it. Thus, our problem formulation may be viewed as an actor/agent choosing a policy (the outer problem) and the adversary then acting to worsen the performance of that strategy (the inner problem). We develop explicit policy gradient formulae for the inner and outer problems and show their efficacy on three prototypical financial problems: robust portfolio allocation, benchmark optimization, and statistical arbitrage.
科研通智能强力驱动
Strongly Powered by AbleSci AI