Abstract
This work studies the performance and the results of the application of Evolutionary Algorithms (EAs) for evolving the decision engine of a program, called in this context agent, which controls the player’s behaviour in an real-time strategy game (RTS). This game was chosen for the Google Artificial Intelligence Challenge in 2011, and simulates battles between teams of ants in different types of maps or mazes. According to the championship rules the agents cannot save information from one game to the next, which makes impossible to implement an EA ‘inside’ the agent, i.e. on game time (or on-line), that is why in this paper we have evolved this engine off-line by means of an EA, used for tuning a set of constants, weights and probabilities which direct the rules. This evolved agent has fought against other successful bots which finished in higher positions in the competition final rank. The results show that, although the best agents are difficult to beat, our simple agent tuned with an EA can outperform agents which have finished 1000 positions above the untrained version.
Access provided by Autonomous University of Puebla. Download to read the full chapter text
Chapter PDF
Similar content being viewed by others
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
References
Buro, M.: Call for AI research in RTS games. In: Proc. AAAI workshop on Challenges in Game AI, pp. 139–141 (2004)
Eiben, A., Smith, J.E.: What is an evolutionary algorithm? In: Rozenberg, G. (ed.) Introduction to Evolutionary Computing, pp. 15–35. Addison Wesley (2005)
Falke-II, W., Ross, P.: Dynamic strategies in a real-time strategy game. In: Cantú-Paz, E., et al. (eds.) GECCO 2003. LNCS, vol. 2724, pp. 1920–1921. Springer, Heidelberg (2003)
Goldberg, D.E.: Genetic Algorithms in search, optimization and machine learning. Addison Wesley (1989)
Google. Google AI Challenge 2011: ANTS (2011), http://aichallenge.org/
Holdum, K.H., Kaysø-Rørdam, C., Østergaard, C.: Google ai challenge 2011: Ants. Jørgen Villadsen, 11 (2011)
Jang, S.H., Yoon, J.W., Cho, S.B.: Optimal strategy selection of non-player character on real time strategy game using a speciated evolutionary algorithm. In: IEEE Symposium on Computational Intelligence and Games, CIG 2009, pp. 75–79. IEEE (2009)
Laird, J.E.: Using a computer game to develop advanced ai. Computer 7(34), 70–75 (2001)
Lidén, L.: Artificial stupidity: The art of intentional mistakes. AI Game Programming Wisdom 2, 41–48 (2004)
Martín, E., Martínez, M., Recio, G., Saez, Y.: Pac-mAnt: Optimization based on ant colonies applied to developing an agent for ms. pac-man. In: 2010 IEEE Conference on Computational Intelligence and Games, CIG 2010, pp. 458–464 (2010)
Mora, A.M., Fernández-Ares, A., Merelo-Guervós, J.-J., García-Sánchez, P.: Dealing with noisy fitness in the design of a RTS game bot. In: Di Chio, C., et al. (eds.) EvoApplications 2012. LNCS, vol. 7248, pp. 234–244. Springer, Heidelberg (2012)
Mora, A.M., Moreno, M.A., Merelo, J.J., Castillo, P.A., García-Arenas, M.I., Laredo, J.L.J.: Evolving the cooperative behaviour in UnrealTM bots. In: Proc. 2010 IEEE Conference on Computational Intelligence and Games, CIG 2010, pp. 241–248 (2010)
Mora, A.M., Fernández-Ares, A., Merelo Guervós, J.J., García-Sánchez, P., Fernandes, C.M.: Effect of noisy fitness in real-time strategy games player behaviour optimisation using evolutionary algorithms. J. Comput. Sci. Technol. 27(5), 1007–1023 (2012)
Onieva, E., Pelta, D.A., Alonso, J., Milans, V., Prez, J.: A modular parametric architecture for the torcs racing engine. In: Proc. 2009 IEEE Symposium on Computational Intelligence and Games, CIG 2009, pp. 256–262 (2009)
Ontañón, S., Mishra, K., Sugandh, N., Ram, A.: Case-based planning and execution for real-time strategy games. In: Weber, R.O., Richter, M.M. (eds.) ICCBR 2007. LNCS (LNAI), vol. 4626, pp. 164–178. Springer, Heidelberg (2007)
Ponsen, M., Munoz-Avila, H., Spronck, P., Aha, D.W.: Automatically generating game tactics through evolutionary learning. AI Magazine 27(3), 75–84 (2006)
Spronck, P., Sprinkhuizen-Kuyper, I., Postma, E.: Improving opponent intelligence through offline evolutionary learning. International Journal of Intelligent Games & Simulation 2(1), 20–27 (2003)
Sweetser, P.: Emergence in games. Game Development (2008)
Togelius, J., Karakovskiy, S., Koutnik, J., Schmidhuber, J.: Super mario evolution. In: IEEE Symposium on Computational Intelligence and Games, CIG 2009, pp. 156–161 (2009)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Carpio, J. et al. (2013). Evolving the Strategies of Agents for the ANTS Game. In: Rojas, I., Joya, G., Cabestany, J. (eds) Advances in Computational Intelligence. IWANN 2013. Lecture Notes in Computer Science, vol 7903. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-38682-4_35
Download citation
DOI: https://doi.org/10.1007/978-3-642-38682-4_35
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-38681-7
Online ISBN: 978-3-642-38682-4
eBook Packages: Computer ScienceComputer Science (R0)