强化学习
操作员(生物学)
选择(遗传算法)
人工智能
计算机科学
符号
数学优化
进化算法
遗传算子
机器学习
过程(计算)
人工神经网络
最优化问题
数学
转录因子
算术
生物化学
基因
操作系统
抑制因子
元优化
化学
作者
Ye Tian,Xiaopeng Li,Haiping Ma,Xingyi Zhang,Kay Chen Tan,Yaochu Jin
标识
DOI:10.1109/tetci.2022.3146882
摘要
Evolutionary algorithms (EAs) have become one of the most effective techniques for multi-objective optimization, where a number of variation operators have been developed to handle the problems with various difficulties. While most EAs use a fixed operator all the time, it is a labor-intensive process to determine the best EA for a new problem. Hence, some recent studies have been dedicated to the adaptive selection of the best operators during the search process. To address the exploration versus exploitation dilemma in operator selection, this paper proposes a novel operator selection method based on reinforcement learning. In the proposed method, the decision variables are regarded as states and the candidate operators are regarded as actions. By using deep neural networks to learn a policy that estimates the $Q$ value of each action given a state, the proposed method can determine the best operator for each parent that maximizes its cumulative improvement. An EA is developed based on the proposed method, which is verified to be more effective than the state-of-the-art ones on challenging multi-objective optimization problems.
科研通智能强力驱动
Strongly Powered by AbleSci AI