强化学习
计算机科学
概率逻辑
可扩展性
马尔可夫决策过程
控制器(灌溉)
数学优化
人工智能
马尔可夫过程
数学
农学
数据库
生物
统计
作者
Amol Yerudkar,Evangelos Chatzaroulas,Carmen Del Vecchio,Sotiris Moschoyiannis
标识
DOI:10.1016/j.ins.2022.11.030
摘要
The rise of reinforcement learning (RL) has guided a new paradigm: unraveling the intervention strategies to control systems with unknown dynamics. Model-free RL provides an exhaustive framework to devise therapeutic methods to alter the regulatory dynamics of gene regulatory networks (GRNs). This paper presents an RL-based technique to control GRNs modeled as probabilistic Boolean control networks (PBCNs). In particular, a double deep- Q network (DD Q N) approach is proposed to address the sampled-data control (SDC) problem of PBCNs, and optimal state feedback controllers are obtained, rendering the PBCNs stabilized at a given equilibrium point. Our approach is based on options , i.e., the temporal abstractions of control actions in the Markov decision processes (MDPs) framework. First, we define options and hierarchical options and give their properties. Then, we introduce multi-time models to compute the optimal policies leveraging the options framework. Furthermore, we present a DD Q N algorithm: i) to concurrently design the feedback controller and the sampling period; ii) wherein the controller intelligently decides the sampled period to update the control actions under the SDC scheme. The presented method is model-free and offers scalability, thereby providing an efficient way to control large-scale PBCNs. Finally, we compare our control policy with state-of-the-art control techniques and validate the presented results.
科研通智能强力驱动
Strongly Powered by AbleSci AI