Extreme Compass and Dynamic Multi-Armed Bandits for Adaptive Operator Selection
Jorge Maturana, Alvaro Fialho, Frédéric Saubion, Marc Schoenauer and Michele Sebag
In: CEC'09: Proceedings of the IEEE International Conference on Evolutionary Computation, May 2009, Trondheim.
The goal of Adaptive Operator Selection (AOS) is the on-line control of the choice of variation operators within Evolutionary Algorithms. The control process is based on two main components, the credit assignment, that defines the reward that will be used to evaluate the quality of an operator after it has been applied, and the operator selection mechanism, that selects one operator based on all operators qualities. Two previously developed AOS methods are combined here: Compass evaluates the performance of operators by considering not only the fitness improvements from parent to offspring, but also the way they modify the diversity of the population, and their execution time; Dynamic Multi-Armed Bandit proposes a selection strategy based on the well-known UCB algorithm, achieving a compromise between exploitation and exploration, while nevertheless quickly adapting to changes. Tests with the proposed method, called ExCoDyMAB, are carried out using several hard instances of the Satisfiability problem (SAT). Results show the good synergetic effect of combining both approaches.