Loading…
Extreme compass and Dynamic Multi-Armed Bandits for Adaptive Operator Selection
The goal of adaptive operator selection is the on-line control of the choice of variation operators within evolutionary algorithms. The control process is based on two main components, the credit assignment, that defines the reward that will be used to evaluate the quality of an operator after it ha...
Saved in:
Main Authors: | , , , , |
---|---|
Format: | Conference Proceeding |
Language: | English |
Subjects: | |
Online Access: | Request full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | The goal of adaptive operator selection is the on-line control of the choice of variation operators within evolutionary algorithms. The control process is based on two main components, the credit assignment, that defines the reward that will be used to evaluate the quality of an operator after it has been applied, and the operator selection mechanism, that selects one operator based on some operators qualities. Two previously developed adaptive operator selection methods are combined here: Compass evaluates the performance of operators by considering not only the fitness improvements from parent to offspring, but also the way they modify the diversity of the population, and their execution time; dynamic multi-armed bandit proposes a selection strategy based on the well-known UCB algorithm, achieving a compromise between exploitation and exploration, while nevertheless quickly adapting to changes. Tests with the proposed method, called ExCoDyMAB, are carried out using several hard instances of the satisfiability problem (SAT). Results show the good synergetic effect of combining both approaches. |
---|---|
ISSN: | 1089-778X 1941-0026 |
DOI: | 10.1109/CEC.2009.4982970 |