集合(抽象数据类型)
班级(哲学)
计算机科学
优势和劣势
领域(数学)
进化算法
考试(生物学)
测试用例
进化计算
可扩展性
试验装置
数学优化
人工智能
机器学习
数学
程序设计语言
回归分析
认识论
纯数学
哲学
生物
数据库
古生物学
作者
Simon Huband,Philip Hingston,Luigi Barone,Lyndon While
标识
DOI:10.1109/tevc.2005.861417
摘要
When attempting to better understand the strengths and weaknesses of an algorithm, it is important to have a strong understanding of the problem at hand. This is true for the field of multiobjective evolutionary algorithms (EAs) as it is for any other field. Many of the multiobjective test problems employed in the EA literature have not been rigorously analyzed, which makes it difficult to draw accurate conclusions about the strengths and weaknesses of the algorithms tested on them. In this paper, we systematically review and analyze many problems from the EA literature, each belonging to the important class of real-valued, unconstrained, multiobjective test problems. To support this, we first introduce a set of test problem criteria, which are in turn supported by a set of definitions. Our analysis of test problems highlights a number of areas requiring attention. Not only are many test problems poorly constructed but also the important class of nonseparable problems, particularly nonseparable multimodal problems, is poorly represented. Motivated by these findings, we present a flexible toolkit for constructing well-designed test problems. We also present empirical results demonstrating how the toolkit can be used to test an optimizer in ways that existing test suites do not
科研通智能强力驱动
Strongly Powered by AbleSci AI