Efficient Parameter Optimization Based on Combination of Direct Global and Local Search Methods
- 331 Downloads
In this paper we focus on direct optimization methods which require nothing but goal function values for orientation (blind search). On the one hand this property ensures robustness and universal applicability. On the other hand direct optimization usually requires a lot of computational effort (goal function evaluations) to ensure optimization success (convergence towards a globally-optimal region of the search space) and an acceptable quality of the optimization result (small approximation error). These fundamental drawbacks of direct optimization are due to the fact that no auxiliary information like derivatives or other problem specific knowledge is exploited to accelerate the optimization process. In order to substantially improve the performance of direct optimization we propose a combination of probabilistic global and deterministic local optimization methods. The resulting combined 2-phase optimization strategy has been proven to be both powerful and efficient. The excellent heuristic properties of this hybrid method allows to use it as the basic component of a multiple-stage optimization strategy. The goal of multiple-stage optimization is to perform a systematic analysis of the most important extreme points of a given optimization problem. This paper presents the structure of our developed optimization algorithms. Beyond that, results of an extensive optimization experiment are presented showing the potentialities but also the limits of our developed methods.
KeywordsGoal Function Problem Dimension Direct Optimization Optimization Stage Optimization Success
Unable to display preview. Download preview PDF.
- E. Aants, J. Korst, Simulated Annealing and Boltzmann Machines, Wiley, 1990.Google Scholar
- D.E. Goldberg Genetic Algorithms in Search, Optimization and Machine Learning, Addison-Wesley, 1989.Google Scholar
- K.-P. Huber, M.R. Berthold, Building Precise Classifiers with Automatic Rule Extraction, Proceedings of the IEEE International Conference on Neural Networks ICNN’95, Perth, Western Australia, Vol. 3, pp. 1263–1268, 1995.Google Scholar
- Z. Michalewicz, Genetic Algorithms d- Data Structures = Evolution Programs, Springer, 1992.Google Scholar
- J.D. Schaffer, R.A. Caruana, L.J. Eshelman, R. Das, A Study of Control Parameters Affecting Online Performance of Genetic Algorithms for Function Optimization, Proceedings of the third International Conference on Genetic Algorithms, June 4–7, George Mason University, pp. 51–60, 1989.Google Scholar
- J. Shekel, Test Functions for Multimodal Search Techniques, in Fifth Annual Princeton Conference on Information Science and Systems, 1971.Google Scholar
- M. Syrjakow, H. Szczekbrka, Optimization of Simulation Models with REMO, Proceedings of the European Simulation Multiconference ESM’94, Barcelona, Spain, June 1–3, pp. 274–281, 1994.Google Scholar
- M. Syrjakow, H. Szczerbicka, Simulation and Optimization of Complex Technical Systems, Proceedings of the 1995 Summer Computer Simulation Conference SCSC’95, Ottawa, Ontario, Canada, July 24–26, pp. 86–95, 1995.Google Scholar
- M. Sykjnkow, H. Szczerbicka, Combination of Direct Global and Local Optimization Methods, Proceedings of the International Conference on Evolutionary Computing ICEC’95, Perth, Western Australia, November 29 - December 1, pp. 326–333, 1995.Google Scholar
- M. Syrjakow, H. Szozerbroka, M.R. Berthold, K.-P. Huger, Acceleration of Direct Model Optimization Methods by Function Approximation,Proceedings of the 8th European Simulation Symposium ESS’96, Genoa, Italy, October 2426, Volume II, pp. 181–186, 1996.Google Scholar