Professional Documents
Culture Documents
Pedro A. Diaz-Gomez
Dean F. Hougen
pdiaz-go@cameron.edu
hougen@ou.edu
ABSTRACT
When an optimization problem is encoded using genetic algorithms, one must address issues of population size, crossover and mutation operators and probabilities, stopping criteria, selection operator and pressure, and fitness function
to be used in order to solve the problem. This paper tests a
relationship between (1) crossover probability, (2) mutation
probability, and (3) selection pressure using two problems.
This relationship is based on the schema theorem proposed
by Holland and reflects the fact that the choice of parameters
and operators for genetic algorithms needs to be problem
specific.
General Terms
Algorithms, Experimentation, Measurement, Performance,
Theory, Verification
Keywords
Performance Analysis, Empirical Study, Evolution Dynamics, Genetic Algorithms, Theory, Working Principles of Evolutionary Computing, Parameter Tuning, Schema Theorem
1.
INTRODUCTION
This paper looks at three genetic algorithm (GA) parameters: (1) crossover probability, (2) mutation probability, and
(3) selection pressure. Through an empirical study based on
Hollands schema theorem [11] we demonstrate that these
2.
Conducting research at the Robotics, Evolution, Adaptation, and Learning Laboratory (REAL Lab), School of Computer Science, University of Oklahoma
Permission to make digital or hard copies of all or part of this work for
personal or classroom use is granted without fee provided that copies are
not made or distributed for profit or commercial advantage and that copies
bear this notice and the full citation on the first page. To copy otherwise, to
republish, to post on servers or to redistribute to lists, requires prior specific
permission and/or a fee.
GECCO09, July 812, 2009, Montral, Qubec, Canada.
Copyright 2009 ACM 978-1-60558-325-9/09/07 ...$5.00.
763
d()
E(, t + 1) 1 pc
sp n(, t)
(1)
l1
d()
sp 1
(2)
1 pc
l1
3.
EXPERIMENTAL APPROACH
3.1
1
l1
1
(3)
pc
d()
sp
Problems Tested
with sp > 1.
Building on this, the equation central to this paper (derived previously [7]), which takes into account mutation
probability, is
l1
1
pc
1
,
(4)
d()
(1 pm h())sp
with pm h() 6= 1, pm > 0, pm 0, and h() > 1. Clearly,
d() and h() depend on which in turn depends on the
problem being encoded for application of the GA, as does
l. This is as it should be, as no one would expect a single set of parameters to be preferable for all optimization
problems [19].
The limits on the parameters for Equation 4 should be
noted, particularly those for pm and sp . In this formulation,
sp = 1 would indicate that the schema in question is no more
likely to be selected than any other schema (that is, there
is no selection in favor of the preferred schema to drive the
evolutionary process), and sp < 1 would indicate selection
against the schema in question, so it is necessary to require
sp > 1.
Similarly, pm = 0 would mean that we run the risk of
never generating chromosomes that match the schema in
question. For example, if we have a specific gene that has
an allele value of 1 for all chromosomes in a population,
then neither cloning a single chromosome nor crossing over
chromosomes with one another can change that value. If a
value of 0 is required for that gene for T , our GA will never
find it using only cloning and crossover. Mutation, then, is
the only operator that can change the value of that gene1 (in
this case to 0) so we cannot allow pm = 0. However, there
is no similar limit on pc . It is true that if pc = 0, then only
mutation can change chromosomes, which may limit a GAs
ability to extract itself from local maxima. However, such
extraction is not impossible, just unlikely, and then only if
3.2
1
This assumes one is not using inversion or any other nonstandard GA operator.
764
3.3
765
4.
3.4
4.1
Problem-Specific Parameters
Data Set
Mean Stdv. Min Q1 Q2 Q3 Max
U (satisfies Eq 4)
87.39 5.27 70 84 88 91 98
V (does not satisfy Eq 4) 85.20 5.16 73 82 85 89 96
0.6
F(x)
3.5
RESULTS
To collect sufficient data for a meaningful statistical analysis, we run the GA 10 times for each triple in U and V ,
recording the best solution found in the last generation for
each run. We then calculate the mean, standard deviation,
maximum, minimum, and quartiles for the sets U and V ,
and perform both Kolmogorov-Smirnov (K-S) and analysis
of variance (ANOVA) tests on the data. Sets U and V are
compared using the K-S test in order to corroborate that
the two sets U and V belong to statistically different distributions and the ANOVA test checks the impact of each
parameter sp , pc , and pm independently, in pairs, and the
three together, on the quality of the solution.
The ANOVA test shows the effect of the independent variables over the dependent variable, as well as if there are
some interactions of the independent variables that affect
the dependent variable. In order to perform the analysis of
variance, it is useful to define a focal independent variable,
a first order moderator independent variable, and a second
order moderator independent variable [12]. It is a common
belief that the crossover operator is the principal factor of
influence in GAs and that the mutation operator is a background operator [11]. So, we propose the crossover operator
as the focal independent variable, the mutation operator as
the first order moderator, and the selection pressure as the
second order moderator. This helps us to analyze the three
way interaction that could exist between crossover, mutation, and selection pressure.
0.5
0.4
0.3
0.2
0.1
0
70
75
80
85
90
95
100
Source
Sum. Sq. d.o.f. Mean Sq.
F
P rob > F
pm
1,115.4
7
159.3
9,225.38
0
pc
57,486.3
10
5,748.6 332,815.17
0
sp
54,477.6
4
13,619.4 788,491.73
0
pm pc
2,605.1
70
37.2
2,154.63
0
pm sp
1,674.1
28
59.8
3,461.53
0
pc sp
1,186.5
40
29.7
1,717.35
0
pm pc sp 6,327.8
280
22.6
1,308.38
0
Error
68.4
3,960
0
T otal
124,941.3 4,399
-
766
4.2
Source
pm
pc
sp
pm pc
pm sp
pc sp
pm pc sp
Error
T otal
0.9
Data Set
Mean Stdv. Min Q1 Q2 Q3 Max
U (satisfies Eq 4)
57.07 7.40
1 53 56 61 79
V (does not satisfy Eq 4) 45.09 11.62 1 39 49 52 71
0.8
0.7
0.6
Data Set
Mean Stdv. Min Q1 Q2 Q3 Max
U (satisfies Eq 4)
57.31 7.61
1 53 56 61 79
V (does not satisfy Eq 4) 44.95 11.88 1 39 49 53 70
0.4
0.3
0.1
0
Source
Sum. Sq. d.o.f. Mean Sq.
F
P rob > F
pm
454,114.9
7
64,873.6 433.92
0
pc
3,294
10
329.4
2.2
0.0151
sp
327,984.6
4
81,996.1 548.45
0
pm pc
15,165.7
70
216.7
1.45
0.009
pm sp
112,379
28
4,013.5 26.85
0
pc sp
10,007.4
40
250.2
1.67
0.0051
pm pc sp
43,609
280
155.7
1.04
0.3101
Error
592,037 3,960
149.5
T otal
1,558,591.6 4,399
-
10
20
30
40
50
60
70
80
0.6
Table 7 shows ANOVA results for case 2. There is a statistically significant impact of each operator alone, pair-wise,
and all the three together on the quality of the solution.
pm
pc
sp
pm pc
pm sp
pc sp
pm pc sp
Error
T otal
0.2
Table 6 shows ANOVA results for case 1. There is a statistically significant impact of each operator alone, and pairwise in the quality of the solution, but not of the three operators together.
Source
0.5
0.5
0.4
0.3
0.2
0.1
0
10
20
30
40
50
60
70
80
767
1
0.9
Satisfied
Not Satisfied
0.8
0.7
90
80
70
0.5
60
Ave. Length
F(x)
0.6
0.4
0.3
50
40
30
20
0
0.004
0.1
0
10
0.2
0.8
0.006
0.6
0.008
0.012
0.014
Mutation Probability
0.016
10
20
30
40
50
60
70
0.4
0.01
80
0.2
0
0.018
Satisfied
Not Satisfied
90
90
80
80
70
70
60
60
Ave. Length
Ave. Length
Satisfied
Not Satisfied
50
40
30
20
10
0
0.004
0.4
0.01
0.012
0.014
Mutation Probability
0.016
40
30
20
0
0.004
0.6
0.008
50
10
0.8
0.006
Crossover Probability
0.8
0.006
0.6
0.008
0.4
0.01
0.012
0.014
Mutation Probability
0.016
0.2
0
0.018
0.2
0
0.018
Satisfied
Not Satisfied
Satisfied
Not Satisfied
90
90
80
80
70
70
60
Ave. Length
60
Ave. Length
Crossover Probability
50
40
30
20
10
0
0.004
Crossover Probability
0.01
0.012
0.014
Mutation Probability
0.016
30
20
0
0.004
0.6
0.4
0.008
40
10
0.8
0.006
50
Crossover Probability
0.8
0.006
0.6
0.008
0.012
0.014
Mutation Probability
0.016
0.2
0
0.018
0.4
0.01
Crossover Probability
0.2
0
0.018
768
5.
DISCUSSION
It has been common practice to use a default set of parameters for GAs, or, in some cases a trial and error procedure is
used [8]. More recently, approaches that systematically test
parameter sets have been proposed [2, 15]. Nonetheless, parameter tuning remains a substantial problem for researchers
in evolutionary computation [16]. The hypothesis tested in
this paper addresses this problem and uses a relationship between pc , pm , and sp as in Equation 4 [7]. Equation 4 shows
that the selection of pc , pm , and sp depends on the problem
itself (l, d(), and h()) and that there is an interrelation
between pc , pm , and sp that could affect the quality of the
solution.
Equation 4 was derived from the schema theorem as described in previous work [7], where it is expected that the
number of target schema T at step t is not going to decrease at step t + 1. It is very unlikely that T is present
in the initial population for all but trivial problems, so the
GA has to build it progressively at each time step. However,
in the building process, destructive factors occur owing to
crossover and mutation, so the probabilities of crossover and
mutation should be such that they minimize that destructive effect to a level that allows quality chromosomes to be
accumulated through selection. This is what the triple parameter hypothesis is proposing with Equation 4. However,
an optimal setting of parameters for GAs is very difficult to
find [8] so the triple parameter hypothesis proposes something more modest: a set U of possible values from which
to draw values for the triple parameters of pc , pm , and sp ,
based on the optimization problem to be solved.
Two problems were chosen to test this hypothesis: the
one-max problem and the snake-in-the-box problem. The
one-max problem is such that the schema under consideration is small (d(T ) = 1) compared to the chromosome
length (we chose l = 100) and the snake-in-the-box-problem
in a 8-dimensional hypercube is such that the schema could
be considered long (d(T ) = 110) relative to the chromosome
length (l = 127) for an 8-dimensional hypercube because a
50-link snake that begins at node 0 and finishes at node 110
was embedded in the lower hypercube. The algorithm needs,
then, to preserve the sub-snake given and look at the 128
nodes of the upper hypercube to complete it.
The set of triples hpc , pm , sp i chosen to perform the tests of
specific hypotheses were such that there is almost no boundary between the points that satisfy Equation 4 (set U ) and
the points that do not satisfy Equation 4 (set V ), which ensures a very challenging test of the hypothesis. The probability of crossover could be set in the full range 0.0 pc 1.0.
The probability of mutation began at 0.004 because for the
case of the one-max problem all the points with pm 0.004
belong to set U (330 in total independent of pc and sp ),
and we wish to have sets of approximately equal size. For
example, for sp = 2, the points that belong to V begin at
pm = 0.010, so the range for pm was between 0.004 and
0.018.
For the one-max problem, as the schema is small, it is
expected that the crossover operator is not destructive. This
fact is so strong that the algorithm performs quite well with
pc high in both sets U and V . However, on average set U
that obeys Equation 4 outperforms set V that does not obey
Equation 4 as is shown in Table 1 in Section 4.1.
Table 2 in Section 4.1 shows the ANOVA results for the
one-max problem. Each operator pc , pm , and sp alone; each
769
ity of the solution. However, there is no statistically significant impact looking at the three independent variables
together. It is expected that if all the individuals already
have T , then they are going to have the same fitness values,
making it harder for sp to moderate the impact of pm in the
relationship between pc and the solution quality.
6.
7.
REFERENCES
[1] T. B
ack. Evolutionary Algorithms in Theory and
Practice. Oxford University Press, 1996.
[2] T. Bartz-Beielstein, C. W. G. Lasarczyk, and
M. Preuss. Sequential parameter optimization. In
Proceedings of the IEEE Congress on Evolutionary
Computation, pages 773780, 2005.
[3] D. S. Bitterman. New lower bounds for the
snake-in-the-box problem: A prolog genetic algorithm
and heuristic search approach. Masters thesis,
University of Georgia, Athens, GA, USA, Dec. 2004.
3
The fitness function defines the optimization problem but
in a practical sense selecting a fitness function is a necessary
step in the implementation of GAs. For example, in the
snake-in-the-box problem, multiple fitness functions try to
find the longest snake, but each one influences the finding of
a better solution [5].
770