Exploiting Sensor Symmetries in Example-based Training for Intelligent Agents (2006)
Intelligent agents in games and simulators often operate in environments subject to symmetric transformations that produce new but equally legitimate environments, such as reflections or rotations of maps. That fact suggests two hypotheses of interest for machine-learning approaches to creating intelligent agents for use in such environments. First, that exploiting symmetric transformations can broaden the range of experience made available to the agents during training, and thus result in improved performance at the task for which they are trained. Second, that exploiting symmetric transformations during training can make the agents' response to environments not seen during training measurably more consistent. In this paper the two hypotheses are evaluated experimentally by exploiting sensor symmetries and potential symmetries of the environment while training intelligent agents for a strategy game. The experiments reveal that when a corpus of human-generated training examples is supplemented with artificial examples generated by means of reflections and rotations, improvement is obtained in both task performance and consistency of behavior.

[Winner of the Best Student Paper award at CIG'06]

See movies of agents in the Legion II strategy game.
In Sushil M. Louis and Graham Kendall, editors, Proceedings of the {IEEE} Symposium on Computational Intelligence and Games, 90-97, Piscataway, NJ, 2006. IEEE.

Bobby D. Bryant Ph.D. Alumni bdbryant [at] cse unr edu
Risto Miikkulainen Faculty risto [at] cs utexas edu