Abstract | ||
---|---|---|
In evolutionary dynamic optimization (EDO), most of the existing studies have assumed that dynamic optimization problems are black boxes. However, for many real-world problems, the dynamic parameters that cause the problems to change are observable. However, determining the utility of these parameters in improving optimization performance has not yet been well studied. In this paper, we propose and compare three strategies for this task: rote learning, fitting data with a feedforward neural network and an ensemble strategy. The main idea of these strategies is to learn the relation between the observable parameters and the optimal solutions and then predict new optima once the environment changes. We also propose a set of test cases representing different kinds of characteristics of real-world problems. In the experiments, the proposed strategies are compared with existing methods that do not use observable parameters, and the results validate our proposed strategies. |
Year | DOI | Venue |
---|---|---|
2020 | 10.1016/j.ins.2019.10.024 | Information Sciences |
Keywords | Field | DocType |
Evolutionary dynamic optimization,gray-box optimization problems,learning,observable parameters,benchmark problems | Rote learning,Feedforward neural network,Observable,Artificial intelligence,Test case,Black box,Optimization problem,Machine learning,Mathematics | Journal |
Volume | ISSN | Citations |
512 | 0020-0255 | 0 |
PageRank | References | Authors |
0.34 | 0 | 4 |
Name | Order | Citations | PageRank |
---|---|---|---|
Tao Zhu | 1 | 82 | 14.36 |
Wenjian Luo | 2 | 356 | 40.95 |
Chenyang Bu | 3 | 47 | 9.18 |
Huansheng Ning | 4 | 847 | 83.48 |