TY - JOUR
T1 - On constructing alternative benchmark suite for evolutionary algorithms
AU - Lou, Yang
AU - Yuen, Shiu Yin
N1 - Publisher Copyright:
© 2018 Elsevier B.V.
PY - 2019/2
Y1 - 2019/2
N2 - Benchmark testing offers performance measurement for an evolutionary algorithm before it is put into applications. In this paper, a systematic method to construct a benchmark test suite is proposed. A set of established algorithms are employed. For each algorithm, a uniquely easy problem instance is generated by evolution. The resulting instances consist of a novel benchmark test suite. Each problem instance is favorable (uniquely easy) to one algorithm only. A hierarchical fitness assignment method, which is based on statistical test results, is designed to generate uniquely easy (or hard) problem instances for an algorithm. Experimental results show that each algorithm performs the best robustly on its uniquely favorable problem. The testing results are repeatable. The distribution of algorithm performance in the suite is unbiased (or uniform), which mimics any subset of real-world problems that is uniformly distributed. The resulting suite offers 1) an alternative benchmark suite to evolutionary algorithms; 2) a novel method of assessing novel algorithms; and 3) meaningful training and testing problems for evolutionary algorithm selectors and portfolios.
AB - Benchmark testing offers performance measurement for an evolutionary algorithm before it is put into applications. In this paper, a systematic method to construct a benchmark test suite is proposed. A set of established algorithms are employed. For each algorithm, a uniquely easy problem instance is generated by evolution. The resulting instances consist of a novel benchmark test suite. Each problem instance is favorable (uniquely easy) to one algorithm only. A hierarchical fitness assignment method, which is based on statistical test results, is designed to generate uniquely easy (or hard) problem instances for an algorithm. Experimental results show that each algorithm performs the best robustly on its uniquely favorable problem. The testing results are repeatable. The distribution of algorithm performance in the suite is unbiased (or uniform), which mimics any subset of real-world problems that is uniformly distributed. The resulting suite offers 1) an alternative benchmark suite to evolutionary algorithms; 2) a novel method of assessing novel algorithms; and 3) meaningful training and testing problems for evolutionary algorithm selectors and portfolios.
KW - Algorithm performance measurement
KW - Evolutionary algorithm
KW - Generating benchmark instance
KW - Hierarchical fitness
KW - Statistical test
UR - http://www.scopus.com/inward/record.url?scp=85046170424&partnerID=8YFLogxK
U2 - 10.1016/j.swevo.2018.04.005
DO - 10.1016/j.swevo.2018.04.005
M3 - Article
AN - SCOPUS:85046170424
SN - 2210-6502
VL - 44
SP - 287
EP - 292
JO - Swarm and Evolutionary Computation
JF - Swarm and Evolutionary Computation
ER -