Evolutionary Computation: Theoretical Issues
Evolutionary Computation: Theoretical Issues
Theoretical Issues
Comparison between evolutionary
algorithms
How to compare Evolutionary Algorithms
Fitness: the
best fitness value found within a certain
number of fitness
How to compare Evolutionary Algorithms
In
fact, computing the fitness function is usually by far the
computationally heaviest part of the algorithm.
How to compare Evolutionary Algorithms
Statistical
tests answer the question: “Considering the
outcome of two algorithms as the sampling of two
random variables, what is the probability (the so-called p-
value) of the ‘null’ hypothesis, given their samplings?”
◦ The lower the p-value, the more the algorithm scoring better
is likely to be actually better than the other.
How to compare Evolutionary Algorithms
Ifdata have a Gaussian distribution, the Student t-test can be
used. It is based on the value of the mean and of the
standard deviation of the two samplings.
Ifdata do not follow a Gaussian distribution, a rank test is
necessary:
◦ Sort ALL results from both distributions
◦ Compute the average rank of data from each distribution and
their standard deviation
The null hypothesis here is that data from the same
distribution have the same average ranking
The p-value is computed using the Wilcoxon test corrected
properly according to the number of samples.
How to compare Evolutionary Algorithms
Then, using
only the successful runs, estimate the average
number of fitness invocations needed to reach the goal.
Example
H = 1##0#1#0 has order o(H) = 4
and defining length d(H) = 7
The schema theorem
The statistical distribution of the number of instances of a
population’s schemata depends mainly on the effects of
genetic operators.
Selection can only modify the statistical distribution of the
pre-existing instances.
Recombination and mutation may create new instances and
cause the extinction of others.
Pd(H,x) denotes the probability that the application of
operator x has a destructive effect on schema H, while Ps(H)
denotes the probability that an instance of schema H is
selected.
The schema theorem
‘No free lunch’ theorem
With reference to evolutionary algorithms, however, one
may ask whether the problems they solve are actually a
sample of the WHOLE problem space or they represent
only a specific region.
In any case, since the result applies to the set of all
possible problems, this does not mean that, considering a
particular problem or class of problems, an optimum
search algorithm may not exist. Consider, for instance,
gradient descent for optimizing unimodal functions, i.e.,
those which do not have any local minima or maxima.
‘No free lunch’ theorem