Abstract
Bayesian Optimization Algorithm (BOA) belongs to the advanced evolutionary algorithms (EA) capable of solving problems with multivariate interactions. However, to attain wide applicability in real-world optimization, BOA needs to be coupled with various efficiency enhancement techniques. A BOA incorporated with a novel entropy-based evaluation relaxation method (eBOA) is developed in this regard. Composed of an on-demand evaluation strategy (ODES) and a sporadic evaluation method, eBOA significantly reduces the number of (fitness) evaluations without imposing any larger population-sizing requirement. Experiments adduce the grounds for its significant improvement in the number of evaluations until reliable convergence. Furthermore, the evaluation relaxation does not negatively affect the scalability performance.
Access provided by Autonomous University of Puebla. Download to read the full chapter text
Chapter PDF
Similar content being viewed by others
Keywords
References
Pelikan, M., Goldberg, D.E., Lobo, F.G.: A Survey of Optimization by Building and Using Probabilistic Models. In: Computational Optimization and Applications, vol. 21, pp. 5–20. Kluwer Academic Publishers, The Netherlands (2002)
Pelikan, M., Goldberg, D.E., Cantu-Paz, E.: BOA: The Bayesian optimization algorithm. In: Proceedings of GECCO 1999, pp. 525–532. Morgan Kaufmann Publishers, San Francisco (1999)
Harik, G.: Linkage learning via probabilistic modeling in the ECGA. Technical Report No. 99010. IlliGAL (1999)
Larrañaga, P., Lozano, J.A.: Estimation of distribution algorithms: A new tool for evolutionary computation. Kluwer Academic Publishers, Boston (2002)
Ahn, C.W., Goldberg, D.E., Ramakrishna, R.S.: Real-coded Bayesian Optimization Algorithm: Bringing the Strength of BOA into the Continuous World. In: Deb, K., et al. (eds.) GECCO 2004. LNCS, vol. 3102, pp. 840–851. Springer, Heidelberg (2004)
Ahn, C.W., Ramakrishna, R.S.: On the Scalability of Real-Coded Bayesian Optimization Algorithm. IEEE Transactions on Evolutionary Computation 12(3), 307–322 (2008)
Pearl, J.: Probabilistic reasoning in intelligent systems: Networks of plausible inference. Morgan Kaufmann, San Mateo (1988)
Sastry, K., Lima, C.F., Goldberg, D.E.: Evaluation Relaxation Using Substructural Information and Linear Estimation. In: Proceedings of GECCO 2006, pp. 419–426. ACM Press, New York (2006)
Pelikan, M., Sastry, K.: Fitness Inheritance in Bayesian Optimization Algorithm. In: Deb, K., et al. (eds.) GECCO 2004. LNCS, vol. 3103, pp. 48–59. Springer, Heidelberg (2004)
Lima, C.F., Pelikan, M., Sastry, K., Butz, M., Goldberg, D.E., Lobo, F.G.: Substructural Neighborhoods for Local Search in the Bayesian Optimization Algorithm. In: Runarsson, T.P., Beyer, H.-G., Burke, E.K., Merelo-Guervós, J.J., Whitley, L.D., Yao, X. (eds.) PPSN 2006. LNCS, vol. 4193, pp. 232–241. Springer, Heidelberg (2006)
Ocenasek, J.: Entropy-Based Convergence Measurement in Discrete Estimation of Distribution Algorithms. Studies in Fuzziness and Soft Computing, vol. 192, pp. 39–50. Springer, Heidelberg (2006)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2010 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Luong, H.N., Nguyen, H.T.T., Ahn, C.W. (2010). Entropy-Based Evaluation Relaxation Strategy for Bayesian Optimization Algorithm. In: García-Pedrajas, N., Herrera, F., Fyfe, C., Benítez, J.M., Ali, M. (eds) Trends in Applied Intelligent Systems. IEA/AIE 2010. Lecture Notes in Computer Science(), vol 6097. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-13025-0_14
Download citation
DOI: https://doi.org/10.1007/978-3-642-13025-0_14
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-13024-3
Online ISBN: 978-3-642-13025-0
eBook Packages: Computer ScienceComputer Science (R0)