Reward-based selection

Reward-based selection is a technique used in evolutionary algorithms for selecting potentially useful solutions for recombination. The probability of being selected for an individual is proportional to the cumulative reward, obtained by the individual. The cumulative reward can be computed as a sum of the individual reward and the reward, inherited from parents.

Description

Reward-based selection can be used within Multi-armed bandit framework for Multi-objective optimization to obtain a better approximation of the Pareto front. [1]

The newborn a'^{(g+1)} and its parents receive a reward r^{(g)}, if a'^{(g+1)} was selected for new population Q^{(g+1)}, otherwise the reward is zero. Several reward definitions are possible:

Reward-based selection can quickly identify the most fruitful directions of search by maximizing the cumulative reward of individuals.

See also

References

  1. Loshchilov, I.; M. Schoenauer; M. Sebag (2011). "Not all parents are equal for MO-CMA-ES" (PDF). Evolutionary Multi-Criterion Optimization 2011 (EMO 2011). Springer Verlag, LNCS 6576. pp. 31–45.
  2. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. (2002). "A fast and elitist multi-objective genetic algorithm: NSGA-II". IEEE Transactions on Evolutionary Computation 6 (2): 182–197. doi:10.1109/4235.996017.
This article is issued from Wikipedia - version of the Friday, July 17, 2015. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.