gas: why do they sometimes not work? n the coding moves the ga to operate on a different search...

6
GAs: why do they sometimes not work? The coding moves the GA to operate on a different search space --- bad coding might deceive the GA or might slow the GA down. Bad coding might also make it very hard for the GA to identify bulding blocks of good solutions, resulting in a somewhat randomized, unfocused search behavior. GA doesn’t use binary strings -> schema theorem does not apply. Population sizes have to be finite. Number of interations have to be finite. Poor handling of constraints. Good solutions are lost or destroyed. Insufficient search space coverage during the search process. too much (-> premature convergence) or not enough selective pressure (- >random search). Unbalanced sampling unbalanced selection of parents. fitness function not well integrated with the selection methods (might need scaling are another selection method). bad random generators.

Upload: earl-mcdowell

Post on 13-Jan-2016

215 views

Category:

Documents


2 download

TRANSCRIPT

Page 1: GAs: why do they sometimes not work? n The coding moves the GA to operate on a different search space --- bad coding might deceive the GA or might slow

GAs: why do they sometimes not work?

The coding moves the GA to operate on a different search space --- bad coding might deceive the GA or might slow the GA down. Bad coding might also make it very hard for the GA to identify bulding blocks of good solutions, resulting in a somewhat randomized, unfocused search behavior.

GA doesn’t use binary strings -> schema theorem does not apply. Population sizes have to be finite. Number of interations have to be finite. Poor handling of constraints. Good solutions are lost or destroyed. Insufficient search space coverage during the search process. too much (-> premature convergence) or not enough selective pressure (->random

search). Unbalanced sampling

– unbalanced selection of parents.– fitness function not well integrated with the selection methods (might need scaling are

another selection method).– bad random generators.

Page 2: GAs: why do they sometimes not work? n The coding moves the GA to operate on a different search space --- bad coding might deceive the GA or might slow

Fitness Function Scaling (cont.)

mineval , maxeval , and aveval denote the maximum, minimum, and average fitness value in a particular population.

Making the search process less randomized by scaling the fitness function:– F(x)=2**(eval(x))

– F(x)=(eval(x)+c)**k with c+mineval >1, k>1 Making the search process more randomized by scaling the fitness function

– F(x)=log(eval(x)+c) with c+mineval >1 The selection strategy that is employed by the GA is very important for fitness

function scaling: fitness functions that work well with one selection strategy might cooperate poorly with others.

By defining fitness functions that are scaled with respect to the average fitness of the population (e.g. -coding [Forrest 1985]); that is, although the average fitness of a population might improve during the evolution process, the selective pressure remains approximately the same, if this approach is successfully applied.

Page 3: GAs: why do they sometimes not work? n The coding moves the GA to operate on a different search space --- bad coding might deceive the GA or might slow

Popular Selection Strategies

• 2 major factors influence the genetic search [Whitley]

• Population Diversity

• Selective Pressure

• Improved Stategies

• Elitist Model (“The best solution never dies.”)

• Expected Value Model (f(v)/fav ) is decreased by subtracting constant or by division through a constant when the chromosome is selected, to reduce stochastic errors in the selection routine.

• Crowding [DeJong] (“a newly generated solution replaces a solution in the population that is similar to it”).

• Ranking Selection

• Tournament Selection (combines fitness-value based selection with ranking selection)

Page 4: GAs: why do they sometimes not work? n The coding moves the GA to operate on a different search space --- bad coding might deceive the GA or might slow

Hybrid Schemes Idea: Combine the GA with other problem solving paradigms,

occasionally taking advantage of already existing problem-specific knowledge that is incorporated into the other problem solving paradigm.

Example: Goldberg’s G-Improvement[1983]:

1. Select one or more strings from the current population.

2. Sweep bit by bit, performing successive one bit changes to the subject strings as long the fitness improves (local search).

3. At the end of the sweep, reinsert the improved strings into the population, and continue the normal GA-process (until the next local search occurs).

GA-Searchrelying on crossover, mutation ,and selection.

Local Search:calculus-based,Greedy or using other techniques

SearchController

Page 5: GAs: why do they sometimes not work? n The coding moves the GA to operate on a different search space --- bad coding might deceive the GA or might slow

Hierarchically Embedded GAs hierarchical approach that employs nested GAs

– Outer GA identifies subregion / reference points / solution framework – Inner GA searches within the subregion / framework provided by the

outer GA. Delta Coding, developed by Whitley et al. applies GA-techniques to two

levels:– level of potential solutions– level of delta changes (slight modifications of a solution)

is also employed by the Dynamic Parameter Encoding Strategy(DPE). Greffenstette employs a Meta-GA/GA setting, in which the Meta-GA learns

the control parameters of a particular GA (similar ideas are also currently explored at Stanford University for various learning alg.).

also related to meta-learning research (-> Stolfo@Columbia) and approaches that employ multi-layered decision making strategies.

Page 6: GAs: why do they sometimes not work? n The coding moves the GA to operate on a different search space --- bad coding might deceive the GA or might slow

Multi-Layered Learning & Optimization Approaches

M1

M1(solves subproblem)Embedded Approaches

M1

M1

M1

Multi-LayeredApproaches

Meta-Decision-Maker

M1

M1 M1 M1

Meta-Learning/Meta-Optimization Tree-like Approaches

Remarks: Meta-decision-makers can use voting, evidence combination or might employ a strategy that was learnt by training the meta-decision maker with meta data. Tree-like approaches employ the same decision making scheme at all levels; however, decisions in intermediate nodes are “navigational, whereas decisions in leafs are final.