Putation 1 Basic Algorithms and Operators by Thomas Back
Putation 1 Basic Algorithms and Operators by Thomas Back
TEAM LRN
Evolutionary Computation 1
Basic Algorithms and Operators
TEAM LRN
EDITORS IN CHIEF
Thomas Back
Associcite Projiessor of Computer Science, Leideri Uni,*ersity, The Netherlund.$;
cind Munuging Director mid Senior Resecirch F e l l o ~ . Center
,
j?)r Applied S y s t e m
Anulysis, Irformcitik Centrirm Dortmund, Germuny
David B Fogel
E.xec*iiti\fe Vice President ctnd c'hiej Scientist, Nuturd Selec-tion ltic,, Oi Jolltr,
Ca11fo rtr in, USA
Zbigniew Michalewicz
Projiissor oj'Computer Science, Univerhity cf North Cw-olinu, Charlotte, USA: cintl
lnstiticte ($Computer scicwce, Polish Acuderny i f Science.$, WLirsctw-,Poland
EDITORIAL BOARD
TEAM LRN
Evolutionary Computation 1
Basic Algorithms and Operators
Edited by
I N S T I T U T E OF PHYSICS PUBLISHING
Bristol and Philadelphia
TEAM LRN
02000 by
PROJECT STAFF
Publisher: Nicki Dennis
Production Editor: Martin Beavis
Production Munuger: Sharon Toop
Assistunt Production Murtuger: Jenny Troyano
Production Controller: Sarah Plenty
Electronic Production Manuger: Tony Cox
TEAM LRN
Contents
Preface
xiii
List of contributors
xvii
Glossary
xxi
TEAM LRN
4
4
6
7
8
9
10
10
18
20
20
21
22
vi
Contents
Principles of genetics
Rayrnorzd C Pntorz
5. I Introduction
5.2 Some fundamental concepts in genetics
5.3 The gene in more detail
5.4 Options for change
5.5 Population thinking
References
23
23
26
27
27
27
33
35
35
38
40
40
41
44
48
51
59
8 Genetic algorithms
Lcirp J Eshelmarl
8.1 Introduction
8.2 Genetic algorithm basics and some variations
8.3 Mutation and crossover
8.4 Representation
8.5 Parallel genetic algorithms
8.6 Conclusion
References
64
9 Evolution strategies
Gunter Ritdnlph
9.1 The archetype of evolution strategies
LRN
9.2 Contemporary evolutionTEAM
strategies
81
59
62
62
64
65
68
75
77
78
78
81
83
Contents
vii
9.3
86
87
10 Evolutionary programming
V Willinin Porto
10. I Introduction
10.2 History
10.3 Current directions
10.4 Future research
References
Further reading
89
89
90
97
1 00
100
102
103
114
13 Hybrid methods
Zbigniewi Micha le w ic;:
References
124
PART 4
103
103
108
109
111
112
I I4
114
117
118
119
120
120
121
122
122
126
REPRESENTATIONS
14 Introduction to representations
Kalyannzoy Deb
14.1 Solutions and representations
14.2 Important representations
14.3 Combined representations
TEAM LRN
References
127
128
130
131
...
Contents
Vlll
15 Binary strings
Thomas Biick
References
132
16 Real-valued vectors
Darkl B Fogel
16.I Object variables
16.2 Object variables and strategy parameters
References
136
17 Permutations
Darrell Whitley
17.1 Introduction
17.2 Mapping integers to permutations
17.3 The inverse of a permutation
17.4 The mapping function
17.5 Matrix representations
17.6 Alternative representations
17.7 Ordering schemata and other metrics
17.8 Operator descriptions and local search
References
139
135
I36
137
138
139
141
141
142
43
45
46
49
49
51
18 Finite-state representations
D m i d B Fogel
18.1 Introduction
18.2 Applications
References
151
152
154
19 Parse trees
Peter J Angeline
References
155
160
21 Other representations
Peter J Angeline and David B Fogel
2 I . 1 Mixed-integer structures
21.2 Introns
2 1.3 Diploid representations
References
163
158
162
163
163
1 64
1 64
PART 5 SELECTION
22 Introduction to selection
Ka lyatimoy De b
166
TEAM LRN
1X
Contents
22.1 Working mechanisms
22.2 Pseudocode
22.3 Theory of selective pressure
References
166
167
170
171
172
24 Tournament selection
181
Tobias Blickle
24.1 Working mechanism
24.2 Parameter settings
24.3 Formal description
24.4 Properties
References
180
181
182
182
183
185
187
25 Rank-based selection
John Grefenstette
25. I
25.2
25.3
25.4
25.5
172
172
175
175
176
Introduction
Linear ranking
Nonlinear ranking
( p , A), ( p + A) and threshold selection
Theory
References
187
188
188
189
190
194
26 Boltzmann selection
Samir W Mahf{)ud
26.1 Introduction
26.2 Simulated annealing
26.3 Working mechanism for parallel recombinative simulated
annealing
26.4 Pseudocode for a common variation of parallel recombinative
simulated annealing
26.5 Parameters and their settings
26.6 Global convergence theory and proofs
References
195
201
TEAM LRN
195
196
196
197
197
199
200
20 1
20 1
Contents
27.3
27.4
27.5
27.6
27.7
27.8
202
202
202
202
203
203
204
205
212
30 Interactive evolution
WolfRnrzg Bnnzhnf
30. I Introduction
30,2 History
30.3 The problem
30.4 The interactive evolution approach
30.5 Difficulties
30.6 Application areas
30.7 Further developments and perspectives
References
Further reading
228
205
206
207
210
21 1
212
213
214
214
218
222
224
225
225
228
228
229
229
23 1
23 1
232
233
234
235
236
xi
Contents
32 Mutation operators
Thomas Biick, Daliid B Fogel, Darrell Whitley mid Peter J Aiigeliiie
32.1 Binary strings
32.2 Real-valued vectors
32.3 Permutations
32.4 Finite-state machines
32.5 Parse trees
32.6 Other representations
References
237
237
239
243
246
248
250
252
33 Recombination
Lnshoii B Bnoker, David B Fogel, Darrell Whitle?; Peter J Aiigeliiie
arzd A E Eiben
33.1 Binary strings
33.2 Real-valued vectors
33.3 Permutations
33.4 Finite-state machines
33.5 Crossover: parse trees
33.6 Other representations
33.7 Multiparent recombination
References
256
34 Other operators
308
256
270
274
284
286
289
289
302
308
317
319
326
329
Index
33 1
TEAM LRN
TEAM LRN
Preface
The original Handbook of Evolutionary Computation (Back et a1 1997) was
designed to fulfil1 the need for a broad-based reference book reflecting the
important role that evolutionary computation plays in a variety of disciplinesranging from the natural sciences and engineering to evolutionary biology and
computer sciences. The basic idea of evolutionary computation, which came
onto the scene in the 195Os, has been to make use of the powerful process of
natural evolution as a problem-solving paradigm, either by simulating it (by
hand or automatically) in a laboratory, or by simulating it on a computer. As
the history of evolutionary computation is the topic of one of the introductory
sections of the Handbook, we will not go into the details here but simply mention
that genetic algorithms, evolution strategies, and evolutionary programming are
the three independently developed mainstream representatives of evolutionary
computation techniques, and genetic programming and classifier systems are the
most prominent derivative methods.
In the 1960s, visionary researchers developed these mainstream methods of
evolutionary computation, namely J H Holland ( 1 962) at Ann Arbor, Michigan,
H J Bremermann (1962) at Berkeley, California, and A S Fraser (1957) at
Canberra, Australia, for genetic algorithms, L J Fogel (1962) at San Diego,
California, for evolutionary programming, and I Rechenberg ( 1965) and H
P Schwefel (1965) at Berlin, Germany, for evolution strategies. The first
generation of books on the topic of evolutionary compuation, written by
several of the pioneers themselves, still gives an impressive demonstration of
the capabilities of evolutionary algorithms, especially if one takes account of
the limited hardware capacity available at that time (see Fogel et a1 (1966),
Rechenberg ( I 973), Holland ( 1975), and Schwefel ( 1977)).
Similar in some ways to other early efforts towards imitating natures
powerful problem-solving tools, such as artificial neural networks and fuzzy
systems, evolutionary algorithms also had to go through a long period of
ignorance and rejection before receiving recognition. The great success that
these methods have had, in extremely complex optimization problems from
various disciplines, has facilitated the undeniable breakthrough of evolutionary
computation as an accepted problem-solving methodology. This breakthrough
is reflected by an exponentially growing number of publications in the field,
and an increasing interest in corresponding conferences and journals. With
these activities, the field now has its own archivable high-quality publications in
which the actual research results are published. The publication of a considerable
amount of application-specific work is, however, widely scattered over different
TEAM LRN
...
Xlll
xiv
Preface
disciplines and their specific conferences and journals, thus reflecting the general
applicability and success of evolutionary computation methods.
The progress in the theory of evolutionary computation methods since
1990 impressively confirms the strengths of these algorithms as well as their
limitations. Research in this field has reached maturity, concerning theoretical
and application aspects, so it becomes important to provide a complete reference
for practitioners, theorists, and teachers in ii variety of disciplines. The
original Hcrridbook of E\vliitioriary Computation was designed to provide such
a reference work. It included complete, clear, and accessible information.
thoroughly describing state-of-the-art evolutionary computation research and
application in a comprehensive style.
These new volumes, based in the original Handbook, but updated, are
designed to provide the material in units suitable for coursework as well as
for individual researchers. The first volume. E\diitionur.~ Computation I :
Basic Afgoritlzms arid Operators, provides the basic information on evolutionary
algorithms. In addition to covering all paradigms of evolutionary computation in
detail and giving an overview of the rationale of evolutionary computation and
of its biological background, this volume also offers an in-depth presentation
of basic elements of evolutionary computation models according to the types
of representations used for typical problem classes (e.g. binary, real-valued,
permutations, finite-state machines, parse trees). Choosing this classification
based on representation, the search operators mutation and recombination
(and others) are straightforwardly grouped according to the semantics of the
data they manipulate. The second volume, Eivlutionary Compiitatiori 2:
Acf\mc.ed Algorithms arid Operutors, provides information on additional topics
of major importance for the design of an evolutionary algorithm, such as
the fitness evaluation, constraint-handling issues, and population structures
(including all aspects of the parallelization of evolutionary algorithms). This
volume also covers some advanced techniques (e.g. parameter control, metaevolutionary approaches, coevolutionary algorithms, etc) and discusses the
efficient implementation of evolutionary algorithms.
Organizational support provided by Institute of Physics Publishing makes it
possible to prepare this second version of the Huricfbook. In particular, we would
like to express our gratitude to our project editor, Robin Rees, who worked with
us on editorial and organizational issues.
References
E\dutiotiury Cotnpiitutioti
Back T, Fogel D B and Michalewicr Z I997 Huti(lhook
TEAM LRN
(Bristol: Institute of Physics Publishing
and New York: Oxford University Press)
References
xv
TEAM LRN
TEAM LRN
List of Contributors
Peter J Angeline (Chapters 19-21, 32, 33)
Senior Scientist, Natural Selection, Inc., Vestal, NY, USA
e-mai 1: angeli ne @ nat ural-selec tion.com
TEAM LRN
xvii
List of Contributors
xviii
David B Fogel (Chapters 1. 4, 6, 16, 18, 20, 21, 27, 32-34, Glossary)
E.rec-uti\v Vice President cind Chief Scientist, Nuturd Selec*tionInc.., Lu Jollu, CA,
USA
e-mail: [email protected]
e-mail: [email protected]
TEAM LRN
List of Contributors
xix
TEAM LRN
TEAM LRN
Glossary
Thornas Back and David B Fogel
Bold text within definitions indicates terms that are also listed elsewhere in this
glossary.
xxi
xxii
G 10ssary
Glossary
xxiii
xxiv
Giossary
cycles between the mates. The cycie crmsover operator preserves absolute
positions of the elements of permutations. (See also Section 33.3.)
Darwinism: The theory of evolution, proposed by Darwin, that evolution
comes about through random variation (mutation) of heritable characteristics, coupled with natural selection, which favors those species for
further survival and evolution that are best adapted to their environmental
conditions. (See also Chapter 4 . )
Deception: Objective functions are called deceptive if the combination of good
building b!ocks by means of recombination !eads to a reduction of fitness
rather than an increase.
Deficiency: A form of mutation that involves a terminal segment loss of
chromosome regions.
Defining length: The defining length of a scheiiia is the maximum distance
between specified positions within the schema. The larger the defining
length of "a schema, the higher becomes its disruption probability by
crossover.
Deletion: A form of mutation that involves an internal segment loss of a
chromosome region.
Deme: An independent subpopulation in the migration model of parallel
evolutionary algorithms.
Diffusion model: The diffusion model denotes a massively parallel
implementation of evolutionary algorithms, where each individual is
realized as a single process being connected to neighboring individuals,
such that a spatial individual structure is assumed. Recombination
and selection an: restricted to the neighborhood of an individual, such
that information is locally preserved and spreads only slowly over the
population.
Dipioid: In diploid organisms, each body ce!! carries two sets of chromosomes;
that is, each chromosome exists in two homologous fGrrns, one of which
is phenotypically realized.
Discrete recombination: Discrete recombination works o n two vectors of
object variables by performing an exchange of the corresponding object
variables with probability one half (other settings of the exchange
probability are in principle possible) (cf uniform crossover). (See cilso
Section 33.2.)
DNA: Deoxyribonucleic acid, a double-stranded macromolecule of helical
structure (comparable to a spiral staircase). Both single strands are linear,
unbranched nucleic acid molecules built up from alternating deoxyribose
(sugar) and phosphate molecules. Each deoxyribose part is coupled to
a nucleotide base, which is responsible for establishing the connection
to the other strand of the DNA, The four nucleotide bases adenine (A),
thymine (T), cytosine ( C ) and guanine (G) are thc alphabet of the genetic
information. The sequences of these bases i n the DNA molecuie determines
TEAM LRN
the building p l a ~of any organism.
Glossary
xxv
xxvi
Glossary
Glossary
xxvii
xxviii
Glossary
Gray code: A binary code for integer values which ensures that adjacent
integers are encoded by binary strings with Hamming distance one.
Gray codes play an important role in the application of canonical genetic
algorithms to parameter optimization problems, because there are certain
situations in which the use of Gray codes may improve the performance of
an evolutionary algorithm.
Hamming distance: For two binary vectors, the Hamming distance is the
number of different positions.
Haploid: Haploid organisms carry one set of genetic information.
Heterozygous: Diploid organisms having different alleles for a given trait.
Hillclimbing strategy: Hillclimbing methods owe their name to the analogy
of their way of searching for a maximum with the intuitive way a sightless
climber might feel his way from a valley up to the peak of a mountain
by steadily moving upwards. These strategies follow a nondecreasing path
to an optimum by a sequence of neighborhood moves. In the case of
multimodal landscapes, hillclimbing locates the optimum closest to the
starting point of its search.
Homologues: Chromosomes of identical structure, but with possibly different
genetic information contents.
Homozygous: Diploid organisms having identical alleles for a given trait.
Hybrid method: Evolutionary algorithms are often combined with classical
optimization techniques such as gradient methods to facilitate an efficient
local search in the final stage of the evolutionary optimization. The
resulting combinations of algorithms are often summarized by the term
hybrid methods.
Implicit parallelism: The concept that each individual solution offers partial
information about sampling from other solutions that contain similar
subsections. Although it was once believed that maximizing implicit
parallelism would increase the efficiency of an evolutionary algorithm,
this notion has been proved false in several different mathematical
developments (See no-free-lunch theorem).
Individual: A single member of a population. In evolutionary algorithms,
an individual contains a chromosome or genome, that usually contains at
least a representation of a possible solution to the problem being tackled
(a single point in the search space). Other information such as certain
strategy parameters and the individual's fitness value are usually also
stored in each individual.
Intelligence: The definition of the term intelligence for the purpose of clarifying
what the essential properties of artificial or computational intelligence
should be turns out to be rather complicated. Rather than taking the usual
anthropocentric view on this, we adopt a definition by D Fogel which
states that intelligence is the capability of a system to adapt its behavior to
TEAM LRN
meet its goals in a range of environments.
This definition also implies that
Glossary
xxix
xxx
Glossary
Migration model: The migration model (often also referred to as the island
model) is one of the basic models of parallelism exploited by evolutionary
algorithm implementations. The population is no longer panmictic,
but distributed in to several independent subpopu I at ions (so-called demes ),
which coexist (typically on different processors, with one subpopulation
per processor) and may mutually exchange information by interdeme
migration. Each of the subpopulations corresponds to a conventional
(i.e. sequential) evolutionary algorithm. Since selection takes place
only locally inside a population, every deme is able to concentrate on
different promising regions of the search space, such that the global
search capabilities of migration models often exceed those of panmictic
populations. The fundamental parameters introduced by the migration
principle are the exchange frequency of information, the number of
individuals to exchange, the selection strategy for the emigrants, and the
replacement strategy for the immigrants.
Monte Carlo algorithm: See uniform random search.
( p . A) strategy: See comma strategy.
( p A) strategy: See plus strategy.
Multiarmed bandit: Classical analysis of schema processing relied on an
analogy to sampling from a number of slot machines (one-armed bandits)
in order to minimize expected losses.
Multimembered evolution strategy: All variants of evolution strategies that
use a parent population size of 1-1 > I and therefore facilitate the utilization
of recombination are summarized under the term multimembered evolution
strategy .
Multiobjective optimization: In multiobjective optimization, the simultaneous
optimization of several, possibly competing, objective functions is required.
The family of solutions to a multiobjective optimization problem is
composed of all those elements of the search space sharing the property that
the corresponding objective vectors cannot be all simultaneously improved.
These solutions are called Pareto optimal.
Multipoint crossover: A crossover operator which uses a predefined number
of uniformly distributed crossover points and exchanges alternating
segments between pairs of crossover points between the parent individuals
(cf one-point crossover).
Mutation: A change of the genetic material, either occurring in the germ path
or in the gametes (generative) or in body cells (somatic). Only generative
mutations affect the offspring. A typical classification of mutations
distinguishes gene mutations (a particular gene is changed), chromosome
mutations (the gene order is changed by translocation or inversion.
or the chromosome number is changed by deficiencies, deletions, or
duplications), and genome mutations (the number of chromosomes or
genomes is changed). In evolutionary algorithms, mutations are either
modeled on the phenotypic TEAM
level LRN
(e.g. by using normally distributed
xxxi
G 10ssary
xxxii
Glossary
tournament selection.
Panmictic population: A mixed population, in which any individual may
be mated with any other individual with a probability that depends only
on fitness. Most conventional evolutionary algorithms have panmictic
populations.
Parse tree: The syntactic structure of any program in computer programming
languages can be represented by a so-called parse tree, where the internal
LRN and leaves of the tree correspond
nodes of the tree correspond toTEAM
operators
xxxiii
G1ossary
xxxiv
Glossary
Punctuated crossover: A crossover operator to explore the potential for selfadaptation of the number of crossover points and their positions. To
achieve this, the vector of object variables is extended by a crossover
mask, where a one bit indicates the position of a crossover point in
the object variable part of the individual. The crossover mask itself is
subject to recombination and mutation to allow for a self-adaptation of the
crossover operator.
Rank-based selection: In rank-based selection methods, the selection
probability of an individual does not depend on its absolute fitness as in
case of proportional selection, but only on its relative fitness in comparison
with the other population members: its rank when all individuals are
ordered in increasing (or decreasing) order of fitness values. (See NISO
Chapter 25. )
Recombination: See crossover.
RNA: Ribonucleic acid. The transcription process in the cell nucleus
generates a copy of the nucleotide sequence on the coding strand of the
DNA. The resulting copy is an RNA molecule, a single-stranded molecule
which carries information by means of the necleotide bases adenine,
cytosine, guanine, and uracil (U) (replacing the thymine in the DNA).
The RNA molecule acts as a messenger that transfers information from the
cell nucleus to the ribosomes, where the protein synthesis takes place.
Scaling function: A scaling function is often used when applying proportional
selection, particularly when needing to treat individuals with non-positive
evaluations. Scaling functions typically employ a linear, logarithmic, or
exponential mapping. (See also Chapter 23.)
Schema: A schema describes a subset of all binary vectors of fixed length
that have similarities at certain positions. A schema is typically specified
by a vector over the alphabet (0, 1, #}. where the ## denotes a wildcard
matching both zero and one.
Schema theorem: A theorem offered to describe the expected number of
instances of a schema that are represented in the next generation of an
evolutionary algorithm when proportional selection is used. Although
once considered to be a fundamental theorem, mathematical results show
that the theorem does not hold in general when iterated over more than one
generation and that it may not hold when individual solutions have noisy
fitness evaluations. Furthermore, the theorem cannot be used to determine
which schemata should be recombined in future generations and has little
or no predictive power.
Segmented crossover: A crossover operator which works similarly to
multipoint crossover, except that the number of crossover points is not
fixed but may vary around an expectation value. This is achieved by a
segment switch rate that specifies the probability that a segment will end
TEAM LRN
at any point in the string.
GI0ssary
xxxv
xxxvi
Glossary
Glossary
xxxvii
bit with a certain probability between the two parent individuals. The
exchange probability typically has a value of one half, but other settings
are possible (cf discrete recombination). (See also Section 33.3.)
Uniform random search: A random search algorithm which samples the
search space by drawing points from a uniform distribution over the search
space. In contrast to evolutionary algorithms, uniform random search does
not update its sampling distribution according to the information gained
from past samples, i.e. it is not a Markov process.
Zygote: A fertilized egg that is always diploid.
TEAM LRN
TEAM LRN
1
Introduction to evolutionary computation
David B Fogel
1.1
Introductory remarks
The term
itself was invented as recently as 1991, and it represents an effort to bring
together researchers who have been following different approaches to simulating
various aspects of evolution. These techniques of genetic algorithms (Chapter 7),
evolution strategies (Chapter 8), and evolutionary programming (Chapter 9) have
one fundamental commonality: they each involve the reproduction, random
variation, competition, and selection of contending individuals in a population.
These form the essential essence of evolution, and once these four processes are
in place, whether in nature or in a computer, evolution is the inevitable outcome
(Atmar 1994). The impetus to simulate evolution on a computer comes from at
least four directions.
1.2 Optimization
Evolution is an optimization process (Mayr 1988, p 104). Darwin ( 1 859, ch 6)
was struck with the organs of extreme perfection that have been evolved, one
such example being the image-forming eye (Atmar 1976). Optimization does not
imply perfection, yet evolution can discover highly precise functional solutions
to particular problems posed by an organisms environment, and even though
the mechanisms that are evolved are often overly elaborate from an engineering
perspective, function is the sole quality that is exposed to natural selection, and
functionality is what is optimized by iterative selection and mutation.
It is quite natural, therefore, to seek to describe evolution in terms of an
algorithm that can be used to solve difficult engineering optimization problems.
The classic techniques of gradient descent, deterministic hill climbing, and
purely random search (with no heredity) have been generally unsatisfactory when
applied to nonlinear optimization problems, especially those with stochastic,
temporal, or chaotic components. But these are the problems that nature has
seemingly solved so very well. Evolution provides inspiration for computing
TEAM LRN
the solutions to problems that have previously appeared intractable. This was a
key foundation for the efforts in evolution strategies (Rechenberg 1965, 1994,
Schwefel 1965, 1995).
1.5 Biology
Rather than attempt to use evolution as a tool to solve a particular engineering
problem, there is a desire to capture the essence of evolution in a computer
simulation and use the simulation to gain new insight into the physics of natural
evolutionary processes (Ray 1991) (see also Chapter 4). Success raises the
possibility of studying alternative biological systems that are merely plausible
images of what life might be like in some way. It also raises the question of what
properties such imagined systems might have in common with life as evolved on
Earth (Langton 1987). Although every model is incomplete, and assessing what
life might be like in other instantiations lies in the realm of pure speculation,
computer simulations under the rubric of artificial life have generated some
LRN
patterns that appear to correspond TEAM
with naturally
occurring phenomena.
Discussion
1.6 Discussion
The ultimate answer to the question why simulate evolution? lies in the lack
of good alternatives. We cannot easily germinate another planet, wait several
millions of years, and assess how life might develop elsewhere. We cannot
easily use classic optimization methods to find global minima in functions when
they are surrounded by local minima. We find that expert systems and other
attempts to mimic human intelligence are often brittle: they are not robust to
changes in the domain of application and are incapable of correctly predicting
future circumstances so as to take appropriate action. In contrast, by successfully
exploiting the use of randomness, or in other words the itsefd use c,furzc.er?ainty.
all possible pathways are open for evolutionary computation (Hofstadter 1995,
p 1 IS). Our challenge is, at least in some important respects, to not allow our
own biases to constrain the potential for evolutionary computation to discover
new solutions to new problems in fascinating and unpredictable ways. However,
as always, the ultimate advancement of the field will come from the careful
abstraction and interpretation of the natural processes that inspire it.
References
Atmar J W 1976 Speculation on the Evolution of Intelligence und its Possible Reuli:ation
in Machine Form Doctoral Dissertation, New Mexico State University
Atmar W 1994 Notes on the simulation of evolution IEEE Truns. Neural NrtMwks NN-5
130-47
Darwin C R 1859 On the Origin of Species by Means of Nutiiral Selection or the
Presenwtion of Failoured Races in the Struggle for L f e (London: Murray)
Fogel D B 1995 Eidutionan Computation: ToMard a Neuv Philosophy cf Machine
Intelligence (Piscataway, NJ: IEEE)
Fogel L J 1962 Autonomous automata Industr. Res. 4 14-9
Fogel L J, Owens A J and Walsh M J 1966 Artijicial Intelligence through Simulated
Evolution (New York: Wiley)
Hofstadter D I995 Fluid Concepts and Creative Analogies: Computer Models of the
Fundamental Mechanisms of Thought (New York: Basic Books)
Holland J H 1975 Adaptation in Natural and Artificial Systems (Ann Arbor, MI:
University of Michigan Press)
Langton C G 1987 Artificial life Art$cial Life ed C G Langton (Reading, MA: AddisonWesley) pp 1 4 7
Mayr E 1988 Touwd a Neu- Philosophy of Biology: 0bser~~ution.s
of an Eidutionist
(Cambridge, MA: Belknap)
Ray T 1991 An approach to the synthesis of life Art$cinl Ltfe I1 ed C G Langton, C
Taylor, J D Farmer and S Rasmussen (Reading, MA: Addison-Wesley) pp 37 1-408
Rechenberg I I965 Cybernetic Solution Path of an Experimental Problem Royal Aircraft
Establishment Library Translation I 122, Farnborough, UK
-I994 Ei,olutionsstrateC:ieS 94 (Stuttgart: Frommann-Holzboog)
Schwefel H-P 1965 Kybernetische Evolution als Strategie der Experimentellen Forschung
in der Striimungstechnik Diploma Thesis, Technical University of Berlin
-1
995 Evolution and Optimum Seeking (New York: Wiley)
TEAM LRN
2
Possible applications of evolutionary
computation
David Beasley
2.1 Introduction
Applications of evolutionary computation (EC) fall into a wide continuum of
areas. For convenience, in this chapter they have been split into five broad
categories:
0
0
0
0
0
planning
design
simulation and identification
control
classification.
Routing
TEAM LRN
Applications in planning
all based at the same depot. A set of customers must each receive one delivery.
Which route should each vehicle take for minimum cost? There are constraints,
for example, on vehicle capacity and delivery times (Blanton and Wainwright
1993, Thangia et a1 1993).
Closely related to this is the transportation problem, in which a single
commodity must be distributed to a number of customers from a number of
depots. Each customer may receive deliveries from one or more depots. What
is the minimum-cost solution? (Michalewicz 1992, 1993).
Planning the path which a robot should take is another route planning
problem. The path must be feasible and safe (i.e. it must be achievable within the
operational constraints of the robot) and there must be no collisions. Examples
include determining the joint motions required to move the gripper of a robot
arm between locations (Parker et a1 1989, Davidor I99 I , McDonnell et a1 1992).
and autonomous vehicle routing (Jakob et a1 1992, Page et a1 1992). In unknown
areas or nonstatic environments, on-line planninghavigating is required, in
which the robot revises its plans as it travels.
2.2.2 Scheduling
Scheduling involves devising a plan to carry out a number of activities over a
period of time, where the activities require resources which are limited, there
are various constraints and there are one or more objectives to be optimized.
Job shop scheduling is a widely studied NP-complete problem (Davis 1985,
Biegel and Davern 1990, Syswerda 1991, Yamada and Nakano 1992). The
scenario is a manufacturing plant, with machines of different types. There are
a number of jobs to be completed, each comprising a set of tasks. Each task
requires a particular type of machine for a particular length of time, and the tasks
for each job must be completed in a given order. What schedule allows all tasks
to be completed with minimum cost? Husbands (1993) has used the additional
biological metaphor of an ecosystem. His method optimizes the sequence of
tasks in each job at the same time as it builds the schedule. In real job shops
the requirements may change while the jobs are being carried out, requiring that
the schedule be replanned (Fang et a1 1993). In the limit, the manufacturing
process runs continuously, so all scheduling must be carried out on-line, as in
a chemical flowshop (Cartwright and Tuson 1994).
Another scheduling problem is to devise a timetable for a set of examinations
(Corne et a1 1994), university lectures (Ling 1992), a staff rota (Easton and
Mansour 1993) or suchlike.
In computing, scheduling problems include efficiently allocating tasks to
processors in a multiprocessor system (Van Driessche and Piessens 1992,
Kidwell 1993, Fogel and Fogel 1996), and devising memory cache replacement
TEAM LRN
policies (Altman et a1 1993).
2.2.3 Packing
Evolutionary algorithms (EAs) have been applied to many packing problems, the
simplest of which is the one-dimensional zero-one knapsack problem. Given a
knapsack of a certain capacity, and a set of items, each with a particular size and
value, find the set of items with maximum value which can be accommodated
in the knapsack. Various real-world problems are of this type: for example, the
allocation of communication channels to customers who are charged at different
rates.
There are various examples of two-dimensional packing problems. When
manufacturing items are cut from sheet materials (e.g. metal or cloth), it is
desirable to find the most compact arrangemerit of pieces, so as to minimize
the amount of scrap (Smith 1985, Fujita et a1 1993). A similar problem arises
in the design of layouts for integrated circuits--how should the subcircuits be
arranged to minimize the total chip area required (Fourman 1985, Cohoon and
Paris 1987, Chan et a1 1991)?
In three dimensions, there are obvious applications in which the best way of
packing objects into a restricted space is required. Juliff (1993) has considered
the problem of packing goods into a truck for delivery.
Janikow and Cai (1992) similarly used EC to estimate statistical functions for
survival analysis in clinical trials. In a similar area, Manela et ul (1993) used
EC to fit spline functions to noisy pharmaceutical fermentation process data.
EC may also be used to identify the sources of airborne pollution, given data
from a number of monitoring points in an urban area-the source apportionment
problem. In electromagnetics, Tanaka et a1 (1993) have applied EC to
determining the two-dimensional current distribution in a conductor, given its
external magnetic field. Away from conventional system identification, an EC
approach has been used to help with identifying criminal suspects. This system
helps witnesses to create a likeness of the suspect, without the need to give an
explicit description.
Applications in classification
10
2.7 Summary
EC has been applied in a vast number of application areas. In some cases it has
advantages over existing computerized techniques. More interestingly, perhaps,
it is being applied to an increasing number of areas in which computers have
not been used before. We can expect to see the number of applications grow
considerably in the future. Comprehensive bibliographies in many different
application areas are listed after the References.
References
Abu Zitar R A and Hassoun M H 1993 Regulator control via genetic search and assisted
reinforcement Proc. 5th Int. Conj: on Genetic Algorithms (Urbunu-Chumpuign, IL,
July 1993) ed S Forrest (San Mateo, CA: Morgan Kaufmann) pp 254-62
Almassy N and Verschure P 1992 Optimizing self-organising control architectures with
genetic algorithms: the interaction between natural selection and ontogenesis
Purullel Problem Solving from Nuture, 2 (Proc. 2nd Int. Conj: on Purullel Problem
Sohping from Nuture, Brussels, 1992) ed R Manner and B Manderick (Amsterdam:
Elsevier) pp 451-60
Altman E R, Agarwal V K and Gao G R 1993 A novel methodology using genetic
algorithms for the design of caches and cache replacement policy Pmc. 5th Irtt.
Conj: on Genetic Algorithms ( Urbunu-Chumpuign, IL, July 1993) ed S Forrest (San
Mateo, CA: Morgan Kaufmann) pp 392-9
Axelrod R 1987 The evolution of strategies in the iterated prisoners dilemma Genetic
Algorithms und Sirnulated Annealing ed L Davis (Boston, MA: Pitman) ch 3, pp 3241
Baba N 1992 Utilization of stochastic automata and genetic algorithms for neural network
learning Purullel Problem Solving from Nuture, 2 (Proc. 2nd Int. Conf on Purullel
Problem Solving frorn Nuture, Brussels, 1992) ed R Manner and B Manderick
(Amsterdam: Elsevier) pp 43 1-40
Bagchi S, Uckun S, Miyabe Y and Kawamura K 1991 Exploring problem-specific
recombination operators for job shop scheduling Proc. 4th lnt. ConJ on Genetic
Algorithms (Sun Diego, CA, July 1991) ed R Belew and L Booker (San Mateo, CA:
Morgan Kaufmann) pp 10-7
Bala J W and Wechsler H 1993 Learning to detect targets using scale-space and genetic
search Proc. 5th lnt. Conj: on Genetic Algorithms ( Urbunu-Champuigtt, IL, July
TEAM LRN
1993) ed S Forrest (San Mateo. CA: Morgan Kaufmann) pp 516-22
References
I1
Biegel J E and Davern J J 1990 Genetic algorithms and job shop scheduling Comput.
Indust. Eng. 19 8 1-9 I
Blanton J L and Wainwright R L 1993 Multiple vehicle routing with time and capacity
constraints Proc. 5th lnt. Cot$ on Genetic Algorithms (Urbana-Champaign, lL, July
1993) ed S Forrest (San Mateo, CA: Morgan Kaufmann) pp 452-9
Booker L 1985 Improving the performance of genetic algorithms in classifier systems
Proc. 1st h i t . Con$ on Genetic Algorithms (Pittsburgh, PA, July 1985) ed J J
Grefenstette (Hillsdale, NJ: Lawrence Erlbaum Associates) pp 80-92
Bramlette M F and Bouchard E E 1991 Genetic algorithms in parametric design of
aircraft Handbook of Genetic Algorithms ed L Davis (New York: Van Nostrand
Reinhold) ch 10, pp 109-23
Cartwright H M and Tuson A L 1994 Genetic algorithms and flowshop scheduling:
towards the development of a real-time process control system Eidutiorzan
Computing (AISB Workshop, Leeds, 1994, Selected Pupers) (Lecture Notes itz
Computer Science 865) ed T C Fogarty (Berlin: Springer) pp 277-90
Chan H, Mazumder P and Shahookar K 1991 Macro-cell and module placement by
genetic adaptive search with bitmap-represented chromosome Ititegrutiotz VLSI J .
12 49-77
Cohoon J P and Paris W D 1987 Genetic placement IEEE Tram. Computer-Aided Design
CAD-6 956-64
Corne D, Ross P and Fang H-L 1994 Fast practical evolutionary timetabling Evolutionary
Computing (AISB Workshop, Leeds, 1994, Selected Papers) (Lecture Notes in
Computer Science 865) ed T C Fogarty (Berlin: Springer) pp 250-63
Cox L A, Davis L and Qiu Y 1991 Dynamic anticipatory routing in circuit-switched
telecommunications networks Handbook of Genetic Algorithms ed L Davis (New
York: Van Nostrand Reinhold) ch 1 I , pp 124-43
Davidor Y 1991 A genetic algorithm applied to robot trajectory generation Handbook
of Genetic Algorithms ed L Davis (New York: Van Nostrand Reinhold) ch 12,
pp 144-65
Davis L 1985 Job shop scheduling with genetic algorithms Proc. 1st hit. Cot$ on Genetic
AIgorithms (Pittsburgh, PA, July 1985) ed J J Grefenstette (Hillsdale, NJ: Lawrence
Erlbaum Associates) pp 1 3 6 4 0
Davis L and Cox A 1993 A genetic algorithm for survivable network design Proc. 5th
Int. Con$ on Genetic Algorithms (Urbana-Champaign, IL, July 1993) ed S Forrest
(San Mateo, CA: Morgan Kaufmann) pp 408-15
DeJong K 1980 Adaptive system design: a genetic approach IEEE Trans. Systems, Man
Cybern. SMC-10 566-74
Easton F F and Mansour N 1993 A distributed genetic algorithm for employee staffing
and scheduling problems Proc. 5th In?. Con5 on Genetic Algorithms (UrbatiaChampaign, IL, July 1993) ed S Forrest (San Mateo, CA: Morgan Kaufmann)
pp 360-67
Etter D M, Hicks M J and Cho K H 1982 Recursive adaptive filter design using
an adaptive genetic algorithm IEEE Int. Con$ on Acoutics, Speech arid Signal
Processing (Piscataway, NJ: IEEE) pp 635-8
Fairley A and Yates D F 1994 Inductive operators and rule repair in a hybrid genetic
learning system: some initial results Evolutionary Computing (AISB Workshop,
Leeds, 1994, Selected Papers) (Lecture Notes in Computer Science 865) ed T C
TEAM LRN
Fogarty (Berlin: Springer) pp 166-79
12
Fang H-L, Ross P and Corne D 1993 A promising genetic algorithm approach to jobshop scheduling, rescheduling and open-shop scheduling problems Proc. 5th lnr.
Conf on Genetic Algorithms (Urbunu-Chumpuigti, IL, July 1993) ed S Forrest (San
Mateo, CA: Morgan Kaufmann) pp 375-82
Feldman D S 1993 Fuzzy network synthesis with genetic algorithms Proc. 5th Int. Cot@
on Genetic Algorithms ( Urbanu-Chumpaign, IL, July 1993) ed S Forrest (San Mateo,
CA: Morgan Kaufmann) pp 3 12-7
Flockton S J and White M 1993 Pole-zero system identification using genetic algorithms
Proc. 5th lnt. Cot$ on Genetic Algorithms ( Urbunu-Chumpuigti, IL, July 1993) ed
S Forrest (San Mateo, CA: Morgan Kaufmann) pp 531-5
Fogarty T C 1994 Co-evolving co-operative populations of rules in learning control
systems Evolutionan Computing (AISB Workshop, Leeds, 1994, Selected Pupers)
(Le(-tureNotes in Computer Science 865) ed T C Fogarty (Berlin: Springer) pp 195209
Fogel D B 1988 An evolutionary approach to the traveling salesman problem Biol.
Cybernet. 6 1 3 9 4 4
-1990
A parallel processing approach to a multiple traveling salesman problem
using evolutionary programming Proc. 4rh Atin. Synip. on Purullel Processing
(Piscataway, NJ: IEEE) pp 318-26
-199 I System Ident$c~ution through Siniuluted Evolution (Needham, MA: Ginn)
-1993a Applying evolutionary programming to selected traveling salesman problems
Cyberner. Sysr. 24 27-36
-1993b Evolving behaviors in the iterated prisoners dilemma E\wlur. Compur. 1
77-97
-1995 E\dutionury Coniputution: T0Mw-d U N e ~ vPhilosophy ($Muchine Intelligenc~e
(Piscataway. NJ: IEEE)
Fogel D B and Fogel L J 1996 Using evolutionary programming to schedule tasks on a
suite of heterogeneous computers Cotput. Operat. Res. 23 527-34
Fogel D B, Fogel L J and Porto V W 1990 Evolving neural networks B i d . Cyhern. 63
387-93
Fogel L J, Owens A J and Walsh M J 1966 Artijicial intelligence Through Simulated
E\diition (New York: Wiley)
Fonseca C M and Fleming P J 1993 Genetic algorithms for multiobjective optimization:
formulation, discussion and generalization Pro(-. 5th lnt. Conf on Genetic
Algorithms ( Urburrci-Chutmpciigti, IL, July 199.1) ed S Forrest (San Mateo, CA:
Morgan Kaufmann) pp 416-23
Fonseca C M, Mendes E M, Fleming P J and Billings S A 1993 Non-linear model
term selection with genetic algorithms Nutiirul Algorithms in Signal Processing
( Workshop, Chelmsfbrd, U K , No\*emher 1993) vol 2 (London: IEE) pp 27/1-27/8
Fourman M P 1985 Compaction of symbolic layout using genetic algorithms Proc. 1st
lnt. Cot$ on Generic Algorithms (Pittsburgh, PA, July 1985) ed J J Grefenstette
(Hillsdale, NJ: Lawrence Erlbaum Associates) pp 14 1-53
Fujita K, Akagi S and Hirokawa N 1993 Hybrid approach for optimal nesting
using genetic algorithm and a local minimization algorithm Ad\unce.s in Design
Automution vol I , DE-65- 1 (ASME) pp 477-84
Furuya H and Haftka R T 1993 Genetic algorithms for placing actuators on space
structures Proc. 5th Irrt. Conj: on Genetic Algorithms ( Urhana-Chantpaigti, IL, July
TEAM LRN
1993) ed S Forrest (San Mateo, CA: Morgan Kaufmann) pp 536-42
References
13
14
References
15
Manela, M., Thornhill N and Campbell J A 1993 Fitting spline functions to noisy data
using a genetic algorithm Proc. 5th Int. Con$ on Genetic Algorithms (UrbanuChampaign, IL, July 1993) ed S Forrest (San Mateo, CA: Morgan Kaufmann)
pp 549-56
McDonnell J R, Andersen B L, Page W C and Pin F G 1992 Mobile manipulator
configuration optimization using evolutionary programming Proc. I st Ann. Conf on
Evolutionary Progrumming ed D B Fogel and W Atmar (La Jolla, CA: Evolutionary
Programming Society) pp 52-62
Melhuish C and Fogarty T C 1994 Applying a restricted mating policy to determine state
space niches using immediate and delayed reinforcement E\dutinnun Compirting
(AISB Workshop, Leeds, 1994, Selected Papers) (Lecture Notes in Computer Science
865) ed T C Fogarty (Berlin: Springer) pp 224-37
Michalewicz Z 1992 Genetic Algorithms + Data Structures = E\wlution ProgrumJ (Berlin:
Springer)
-1993
A hierarchy of evolution programs: an experimental study Eidut. Conrput. 1
5 1-76
Miller G F, Todd P M and Hegde S U 1989 Designing neural networks using genetic
algorithms. Proc. 3rd Int. Conj on Genetic Algoritlims (Fairfar. VA, June 1989) ed
J D Schaffer (San Mateo, CA: Morgan Kaufmann) pp 379-84
Miihlenbein H I989 Parallel genetic algorithms, population genetics and combinatorial
optimization Proc. 3rd hit. Con5 on Genetic Algorithms (Fuirjiar, VA, June 1989)
ed J D Schaffer (San Mateo, CA: Morgan Kaufmann) pp 4 16-2 I
Namibar R and Mars P 1993 Adaptive IIR filtering using natural algorithms Nuturul
Algorithms in Signal Processing (Workshop, Chelm.?ford.UK, Noiwnber 1993) vol
2 (London: IEE) pp 20/1-20/10
Oliver J R 1993 Discovering individual decision rules: an application of genetic
algorithms Proc.. 5th lnt. Con5 on Genetic Algorithms ( Urbana-Chanil~aiRri,
fL,J i c l ~
1993) ed S Forrest (San Mateo, CA: Morgan Kaufmann) pp 216-22
Oliver I M, Smith D J and Holland J R C 1987 A study of permutation crossover operators
on the travelling salesman problem Proc. 2nd Int. Conf on Genetic Algorithms
(Cambridge, MA, 1987) ed J J Grefenstette (Hillsdale, NJ: Erlbaum) pp 224-30
Page W C, McDonnell J R and Anderson B 1992 An evolutionary programming
approach to multi-dimensional path planning Proc.. I st Ann. Conf on E\wlurionur>
Programming ed D B Fogel and W Atmar (La Jolla, CA: Evolutionary Programming
Society) pp 63-70
Parker J K, Goldberg D E and Khoogar A R 1989 Inverse kinematics of redundant robots
using genetic algorithms Proc. Int. Con6 on Robotics und Automation (Scottsdde,
AZ, 1989) vol 1 (Los Alamitos: IEEE Computer Society Press) pp 271-6
Pate1 M J and Dorigo M 1994 Adaptive learning of a robot arm Evolutionuv Computing
(AISB Workshop, Leeds, 1994, Selected Papers) (Lecture Notes in Computer Science
865) ed T C Fogarty (Berlin: Springer) pp 180-94
Pipe A G and Carse B 1994 A comparison between two architectures for searching and
learning in maze problems Evolutionary Computing (AISB Workshop, Leeds, 1994,
Selected Papers) (Lecture Notes in Computer Scierzc-e865) ed T C Fogarty (Berlin:
Springer) pp 2 3 8 4 9
Polani D and Uthmann T 1992 Adaptation of Kohonen feature map topologies by genetic
TEAM
algorithms Parallel Problem Solving
fromLRN
Nature, 2 (Proc. 2nd lnt. Cot$ on Purullrl
16
References
17
Spencer G F 1993 Automatic generation of programs for crawling and walking Proc. 5th
lnt. Con$ on Genetic Algorithms (Urbana-Champaign, IL, July 1993) ed S Forrest
(San Mateo, CA: Morgan Kaufmann) p 654
Spittle M C and Horrocks D H 1993 Genetic algorithms and reduced complexity artificial
neural networks Natural Algorithms in Signal Processing (Workshop, Chelmsford,
UK, November 1993) vol I (London: IEE) pp 811-819
Suckley D 1991 Genetic algorithm in the design of FIR filters IEE Proc. G 138 234-8
Syswerda G 1991 Schedule optimization using genetic algorithms Handbook qf Generic
Algorithms ed L Davis(New York: Van Nostrand Reinhold) ch 21, pp 3 3 2 4 9
Tackett W A 1993 Genetic programming for feature discovery and image discrimination
Proc. 5th Itit. Cot$ on Genetic Algorithms (Urba,icr-Chainl,aigti,IL, July 1993) ed
S Forrest (San Mateo, CA: Morgan Kaufmann) pp 303-9
Tanaka, Y., Ishiguro A and Uchikawa Y 1993 A genetic algorithms application to inverse
problems in electromagnetics Proc. 5th Itit. Con$ on Genetic Algorithms ( UrbunaChampaign, IL, July 1993) ed S Forrest (San Mateo, CA: Morgan Kaufmann) p 656
Thangia S R, Vinayagamoorthy R and Gubbi A V 1993 Vehicle routing with time
deadlines using genetic and local algorithms Proc. 5th hit. Conf on Genetic
Algorithms (Urbana-Champaign, IL, July 1993) ed S Forrest (San Mateo, CA:
Morgan Kaufmann) pp 506-13
Unger R and Moult J 1993 A genetic algorithm for 3D protein folding simulations
Proc. 5th Int. Con& on Genetic Algorithms ( Urbana-CIiunipaigri, IL. July 1993) ed
S Forrest (San Mateo, CA: Morgan Kaufmann) pp 581-8
Van Driessche R and Piessens R 1992 Load balancing with genetic algorithms Pardlel
Problem Solving from Nature, 2 (Proc. 2nd lnt. Cor$ on Parallel Problem Solling
from Nature, Brussels, 1992) ed R Manner and B Manderick (Amsterdam: Elsevier)
pp 341-50
Verhoeven M G A, Aarts E H L, van de Sluis E and Vaessens R J M 1992 Parallel local
search and the travelling salesman problem Parallel Problem Solling from Nurure,
2 (Proc. 2nd Itit. Con5 on Parallel Problem Solving from Nature, Brussels, 1992)
ed R Manner and B Manderick (Amsterdam: Elsevier) pp 543-52
Watabe H and Okino N 1993 A study on genetic shape design Pro(.. 5th Int. Conf on
Genetic Algorithms (Urbana-Champaign, IL, July 1993) ed S Forrest (San Mateo,
CA: Morgan Kaufmann) pp 445-50
White M and Flockton S 1993 A comparative study of natural algorithms for adaptive
IIR filtering Natural Algorithms in Signal Processing (Workshop, Chelmsford, UK,
November 1993) vol 2 (London: IEE) pp 22/1-22/8
Whitley D, Starkweather T and Fuquay D 1989 Scheduling problems and travelling
salesmen: the genetic edge recombination operator Proc. 3rd h i t . Cot$ on Genetic
Algorithms (Fairfim, VA, June 1989) ed J D Schaffer (San Mateo, CA: Morgan
Kaufmann) pp 1 3 3 4 0
Wicks T and Lawson S 1993 Genetic algorithm design of wave digital filters with
a restricted coefficient set Natural Algorithms in Signal Processing (Workshop,
Chelmsford, UK, November 1993) vol 1 (London: IEE) pp 17/1-17/7
Wilson P B and Macleod M D 1993 Low implementation cost IIR digital filter
design using genetic algorithms Natural Algorithms in Signcil Processing (Workshop,
Chelmsford, UK, November 1993) vol 1 (London: IEE) pp 4 / 1 4 / 8
Wilson S W I987 Hierarchical credit allocation in a classifier system Genetic Algorirhms
TEAM
LRN MA: Pitman) ch 8, pp 104-15
(Boston,
and Simulated Annealing ed L Davis
18
Further reading
This article has provided only a glimpse into the range of applications for
evolutionary computing. A series of comprehensive bibliographies has been
produced by J T Alander of the Department of Information Technology and
Production Economics, University of Vaasa, as listed below.
Art and Music: Indexed Bibliography of' Genctic Algorithms in Art and Music
Report 94- 1 -ART (ftp.uwasa.fi/cs/report94- I/gaARTbib.ps.Z)
Chemistry and Physics: Indexed Bibliogruphy of Genetic Algorithms in
Chemistn and Physics Report 94- I -CHEMPHYS (ftp.uwasa.fi/cs/report941/gaCHEMPHY Sbib.ps.Z)
Control: bzdesecl Bihliogruphy of Genetic Algorithms in Control. Report 94- I CONTROL (ftp.uwasa.fi/cs/report94- I /gaCONTKOLbib.ps.Z)
1 . Computer Aided Design: Indexed Bibliography of Genetic Algorithms in Computrr
Aided Design Report 94- I -CAD (ftp.uwasa.fi/cs/report94- 1 /gaCADbib.ps.Z)
3. Economics: Indexed Bihliogruphy of Genetic Algorithms in Economics Report 941 -ECO ( ftp.uwasa. fi/cs/report94- 1/gaECObi b.ps.Z)
4. Electronics and VLSI Design and Testing: Inde=ced Bibliography of Genetic
Algorithms in Electronics mid VLSI Design und Testing Report 94- I -VLSI
(ftp.uwasa.fi/cs/report94- I /gaVLSIbib.ps.Z)
7. Logistics: Indexed Bibliography of Genetic Algorithms in Logi.stic.s Report 94- 1 TEAM1 /gaLOGI
LRN
LOGISTICS (ftp.uwasa. fi/cs/report94ST1CSbib.ps.Z)
19
Further reading
TEAM LRN
3
Advantages (and disadvantages) of
evolutionary computation over other
approaches
Ham-Paul Schwefel
TEAM LRN
Conclusions
21
one(s) with respect to the additional test problem. This game could in principle
be played ad infiniturn.
A better means of clarifying the scene ought to result from theory. This
should clearly define the domain of applicability of each algorithm by presenting
convergence proofs and efficiency results. Unfortunately. however, it is possible
to prove abilities of algorithms only by simplifying them as well as the situations
to which they are confronted. The huge remainder of questions must be
answered by means of (always limited) test series, and even that cannot tell
much about an actual real-world problem-solving situation with yet unanalyzed
features, that is, the normal case in applications.
Again unfortunately, there does not exist an agreed-upon test problem
catalogue to evaluate old as well as new algorithms in a concise way. It is
doubtful whether such a test bed will ever be agreed upon, but efforts in that
direction would be worthwhile.
3.2 Conclusions
Finally, what are the truths and consequences? First, there will always remain a
dichotomy between efficiency and general applicability, between reliability and
effort of problem-solving, especially optimum-seeking, algorithms. Any specific
knowledge about the situation at hand may be used to specify an adequate
specific solution algorithm, the optimal situation being that one knows the
solution in advance. On the other hand, there cannot exist one method that solves
all problems effectively as well as efficiently. These goals are contradictory.
If there is already a traditional method that solves a given problem, EAs
should not be used. They cannot do it better or with less computational effort.
In particular, they do not offer an escape from the curse of dimensionality-the
often quadratic, cubic, or otherwise polynomial increase in instructions used as
the number of decision variables is increased, arising, for example, from matrix
manipulation.
To develop a new solution method suitable for a problem at hand may be
a nice challenge to a theoretician, who will afterwards get some merit for his
effort, but from the application point of view the time for developing the new
technique has to be added to the computer time invested. In that respect, a
nonspecialized, robust procedure (and EAs belong to this class) may be, and
often proves to be, worthwhile.
A warning should be given about a common practice-the linearization or
other decomplexification of the situation in order to make a traditional method
applicable. Even a guaranteed globally optimal solution for the simplified task
may be a long way off and thus greatly inferior to an approximate solution to
the real problem.
The best one can say about EAs, therefore, is that they present a
methodological framework that is easy to understand and handle, and is either
usable as a black-box method orTEAM
open LRN
to the incorporation of new or old
References
Schwefel H-P 1995 Eidution cind Optimum Seeking (New York: Wiley)
Wolpert D H and Macready W G 1996 N o Free Lunch Theorem.sJi)r Seurch Technical
Report SFI-TR-95-02-010 Santa Fe Institute
TEAM LRN
4
Principles of evolutionary processes
David B Fogel
4.1
Overview
The most widely accepted collection of evolutionary theories is the neoDarwinian paradigm. These arguments assert that the vast majority of the
history of life can be fully accounted for by physical processes operating on
and within populations and species (Hoffman 1989, p 39). These processes
are reproduction, mutation, competition, and selection. Reproduction is an
obvious property of extant species. Further, species have such great reproductive
potential that their population size would increase at an exponential rate if
all individuals of the species were to reproduce successfully (Malthus 1826,
Mayr 1982, p 479). Reproduction is accomplished through the transfer of an
individuals genetic program (either asexually or sexually) to progeny. Mutation,
in a positively entropic system, is guaranteed, in that replication errors during
information transfer will necessarily occur. Competition is a consequence of
expanding populations in a finite resource space. Selection is the inevitable
result of competitive replication as species fill the available space. Evolution
becomes the inescapable result of interacting basic physical statistical processes
(Huxley 1963, Wooldridge 1968, Atmar 1979).
Individuals and species can be viewed as a duality of their genetic program,
the genotype (Section 5.2), and their expressed behavioral traits, the phenofype.
The genotype provides a mechanism for the storage of experiential evidence,
of historically acquired information. Unfortunately, the results of genetic
variations are generally unpredictable due to the universal effects of pleiotropy
and polygeny (figure 4.1) (Mayr 1959, 1963, 1982, 1988, Wright 1931, 1960,
Simpson 1949, p 224, Dobzhansky 1970, Stanley 1975, Dawkins 1986).
Pleiotropy is the effect that a single gene may simultaneously affect several
phenotypic traits. Polygeny is the effect that a single phenotypic characteristic
may be determined by the simultaneous interaction of many genes. There are no
one-gene, one-trait relationships in naturally evolved systems. The phenotype
varies as a complex, nonlinear function of the interaction between underlying
genetic structures and current environmental conditions. Very different genetic
TEAM LRN
23
24
Figure 4.1. Pleiotropy is the effect that a single gene may simultaneously affect several
phenotypic traits. Polygeny is the effect that a single phenotypic characteristic may
be determined by the simultaneous interaction of many genes. These one-to-many and
many-to-one mappings are pervasive in natural systems. As a result, even small changes
to a single gene may induce a raft of behavioral changes in the individual (after Mayr
1963 ).
structures may code for equivalent behaviors, just as diverse computer programs
can generate similar functions.
Selection directly acts only on the expressed behaviors of individuals and
species (Mayr 1988, pp 477-8). Wright (1932) offered the concept of adaptive
topography to describe the fitness of individuals and species (minimally, isolated
reproductive populations termed demes). A population of genotypes maps to
respective phenotypes (sensir Lewontin 1974), which are in turn mapped onto
the adaptive topography (figure 4.2). Each peak corresponds to an optimized
collection of phenotypes, and thus to one of more sets of optimized genotypes.
Evolution probabilistically proceeds up the slopes of the topography toward
peaks as selection culls inappropriate phenotypic variants.
Others (Atmar 1979, Raven and Johnson 1986, pp 400-1) have suggested
that i t is more appropriate to view the adaptive landscape from an inverted
position. The peaks become troughs, 'minimized prediction error entropy wells'
(Atmar 1979). Searching for peaks depicts evolution as a slowly advancing,
tedious, uncertain process. Moreover, there appears to be a certain fragility to
an evolving phyletic line; an optirnized population might be expected to quickly
fall of the peak under slight perturbations. The inverted topography leaves an
altogether different impression. Populations advance rapidly down the walls of
TEAM
the error troughs until their cohesive
set LRN
of interrelated behaviors is optimized,
Overview
25
26
References
Atmar W I979 The inevitability of evolutionary invention, unpublished manuscript
Dawkins R 1986 The Blind Whtchmuker (Oxford: Clarendon)
Dobzhansky T I970 Genetita of the Evolutionury Processes (New York: Columbia
University Press)
Hoffman A 1989 Arguments on Evolution: a Puleontologist's Perspectitte (New York:
Oxford University Press)
Huxley J 1963 The evolutionary process Evolution UJ (1 Process ed J Huxley, A C Hardy
and E B Ford (New York: Collier) pp 9-33
Lewontin R C 1974 The Genetic Busis of E v olutionq Chunge (New York: Columbia
University Press)
Malthus T R 1826 An Essuy on the Principle of Population, a.\ it Aflects the Future
Impro\wnent of Socieh 6th edn (London: Murray)
Mayr E 1959 Where are we? Cold Spring Hurbor Symp. Quunt. Biol. 24 409-40
-1963 Anirnul Species und E\dution (Cambridge, MA: Belknap)
-I982 The Gro,rvth of Biologicd Thought: Diversity, Evolution and Inheritunce
(Cambridge, MA: Belknap)
-1988 Towurd N New- Philosophy of Biology: Observutions of an Evolutionist
(Cambridge, MA: Belknap)
Raven P H and Johnson G B 1986 Biology (St Louis, MO: Times Mirror)
Simpson G G 1949 The Mecining of Evolution: U Study of the History of Life and its
SigniJicuncejbr Mun (New Haven, CT: Yale University Press)
Stanley S M 1975 A theory of evolution above the species level Proc. Nut1 Amd. Sci.
USA 72 646-50
ul Busis of Intelligent L f e (New
Wooldridge D E 1968 The Mechanicul Mun: the Ph
York: McGraw-Hill)
Wright S 1931 Evolution in Mendelian populations Genetics 16 97-1 59
-1932 The roles of mutation, inbreeding, crossbreeding, and selection in evolution
Proc. 6th Int. Congr. on Genetic5 (Ithucu, N Y ) vol 1, pp 356-66
-1960
The evolution of life, panel discussion Eivlution After Dumin: Issues in
Eidution vol 3 , ed S Tax and C Callender (Chicago, IL: University of Chicago
Press)
TEAM LRN
5
Principles of genetics
Raymond C Paton
5.1 Introduction
The material covers a number of key areas which are necessary to understanding
the nature of the evolutionary process. We begin by looking at some basic ideas
of heredity and how variation occurs in interbreeding populations. From here
we look at the gene in more detail and then consider how i t can undergo change.
The next section looks at aspects of population thinking needed to appreciate
selection. This is crucial to an appreciation of Darwinian mechanisms of
evolution. The chapter concludes with selected references to further information.
In order to keep this contribution within its size limits, the material is primarily
about the biology of higher plants and animals.
27
28
Principles of genetics
rule is not universally true when it comes to the distribution of sex chromosomes.
Human diploid cells contain 46 chromosomes of which there are 22 pairs and
an additional two sex chromosomes. Sex is determined by one pair (called
the sex chromosomes); female is X and male is Y. A female human has the
sex chromosome genotype of XX and a male is XY. The inheritance of sex is
summarized in figure 5.2. The members of a pair of nonsex chromosomes are
said to be homologous (this is also true for XX genotypes whereas XY are not
homologous).
29
of particular traits in peas. For example, he took plants that had wrinkled
seeds and plants that had round seeds and bred them with plants of the same
phenotype (i.e. observable appearance), so wrinkled were bred with wrinkled and
round were bred with round. He continued this over a number of generations
until round always produced round offspring and wrinkled, wrinkled. These
are called pure breeding plants. He then cross-fertilized the plants by breeding
rounds with wrinkles. The subsequent generation (called the FI hybrids) was
all round. Then Mendel crossed the F1 hybrids with each other and found that
the next generation, the F2 hybrids, had round and wrinkled plants in the ratio
of 3 (round) : I (wrinkled).
Mendel did this kind of experiment with a number of pea characteristics
such as:
color of cotyledons
color of flowers
color of seeds
length of stem
yellow or green
red or white
graybrown or white
tall or dwarf.
In each case he found that the the Fl hybrids were always of one form and
the two forms reappeared in the F2. Mendel called the form which appeared in
the F1 generation dominant and the form which reappeared in the F2 recessive
(for the full text of Mendels experiments see an older genetics book, such as
that by Sinnott et nl (1958)).
A modern interpretation of inheritance depends upon a proper understanding
of the nature of a gene and how the gene is expressed in the phenotype. The
nature of a gene is quite complex as we shall see later (see also Alberts et nl
1989, Lewin 1990, Futuyma 1986). For now we shall take it to be the functional
unit of inheritance. An allele (allelomorph) is one of several forms of a gene
occupying a given locus (location) on a chromosome. Originally related to pairs
of contrasting characteristics (see examples above), the idea of observable unit
characters was introduced to genetics around the turn of this century by such
workers as Bateson, de Vries, and Correns (see Darden 1991). The concept of
a gene has tended to replace allele in general usage although the two terms are
not the same.
How can the results of Mendels experiments be interpreted? We know
that each parent plant provides half the chromosome complement found in its
offspring and that chromosomes in the diploid cells are in pairs of homologues.
In the pea experiments pure breeding parents had homologous chromosomes
which were identical for a particular gene; we say they are homozygous for
a particular gene. The pure breeding plants were produced through selffertilization and by selecting those offspring of the desired phenotype. As round
was dominant to wrinkled we say that the round form of the gene is R (big
r) and the wrinkled r (little r). Figure 5.3 summarizes the cross of a pure
breeding round (RR) with a pure breeding
wrinkled (rr).
TEAM LRN
We see the appearance of the heterozygote (in this case Rr) in the F1
generation. This is phenotypically the same as the dominant phenotype but
genotypically contains both a dominant and a recessive form of the particular
gene under study. Thus when the heterozygotes are randomly crossed with
each other the phenotype ratio is three dominant : one recessive. This is called
the rnonohybrid ratio (i.e. for one allele). We see in Mendels experiments
the independent segregation of alleles during breeding and their subsequent
independent as sortmen t in offspring .
In the case of two genes we find more phenotypes and genotypes appearing.
Consider what happens when pure breeding homozygotes for round yellow seeds
(RRYY) are bred with pure breeding homozygotes for wrinkled green seeds
(rryy). On being crossed we end up with heterozygotes with a genotype of
RrYy and phenotype of round yellow seeds. We have seen that the genes
segregate independently during meiosis so we have the combinations shown in
figure 5.4.
R r Y y
genes segregate
independently
RY
Ry
rY ry
Thus the gametes of the heterozygote can be of four kinds though we assume
that each form can occur with equal frequency. We may examine the possible
combinations of gametes for the next generation by producing a contingency
TEAM LRN
table for possible gamete combinations.
These are shown in figure 5.5.
31
32
Principles of genetics
33
34
Principles of genetics
translated into protein. The translation process converts the mRNA code into a
protein sequence via another form of RNA called transfer RNA (tRNA). In this
way, genes are transcribed so that mRNA may be produced, from which protein
molecules (typically the workhorses and structural molecules of a cell) can be
formed. This flow of information is generally unidirectional. (For more details
on this topic the reader should consult a molecular biology text and look at the
central dogma of molecular biology, see e.g. Lewin 1990, Alberts et a1 1989.)
Figure 5.1 1 provides a simplified view of the anatomy of a structural gene,
that is, one which codes for a protein or RNA.
That part of the gene which ultimately codes for protein or RNA is preceded
upstream by three stretches of code. The enhancer facilitates the operation of
the promoter region. which is where RNA polymerase is bound to the gene in
order to initiate transcription. The operator is the site where transcription can
be halted by the presence of a repressor protein. Exons are expressed in the
final gene product (e.g. the protein molecule) whereas introns are transcribed
but are removed from the transcript leaving the fragments of exon material to
be spliced. One stretch of DNA may consist of several overlapping genes. For
example, the introns in one gene may be the exons in another (Lewin 1990).
The terminator is the postexon region of the gene which causes transcription
to be terminated. Thus a biological gene contains not only code to be read
but also coded instructions on how it should be read and what should be read.
Genes are highly organized. An operon system is located on one chromosome
and consists of a regulator gene and a number of contiguous structural genes
which share the same promoter and terminator and code for enzymes which
are involved in specific metabolic pathways (the classical example is the Lac
operon, see figure 5.12).
Operons can be grouped together into higher-order (hierarchical) regulatory
genetic systems (Neidhart et crl 1990). For example, a number of operons
from different chromosomes may be regulated by a single gene known as a
regulon. These higher-order systems provide a great challenge for change in a
genome. Modification of the higher-order gene can have profound effects on
TEAM
the expression of structural genes that
areLRN
under its influence.
35
5.4
We have already seen how sexual reproduction can mix up the genes which
are incorporated in a gamete through the random reassortment of paternal
and maternal chromosomes and through crossing over and recombination.
Effectively though. the gamete acquires a subset of the same genes as the
diploid gamete-producing cells; they are just mixed up. Clearly, any zygote that
is produced will have a mixture of genes and (possibly) some chromosomes
which have both paternal and maternal genes.
There are other mechanisms of change which alter the genes themselves
or change the number of genes present in a genome. We shall describe a
mutation as any change in the sequence of genomic DNA. Gene mutations
are of two types: point mutation. in which a single base is changed, and
frameshift mutation, in which one or more bases (but not a multiple of three)
are inserted or deleted. This changes the frame in which triplets are transcribed
into RNA and ultimately translated into protein. In addition some genes are
able to become transposed elsewhere in a genome. They jump about and
are called transposons. Chromosome changes can be caused by deletion (loss
of a section), duplication (the section is repeated), inversion (the section is in
the reverse order), and translocation (the section has been relocated elsewhere).
There are also changes at the genome level. Ploidy is the term used to describe
multiples of a chromosome complement such as haploid ( n ) , diploid (212), and
tetraploid (4n). A good example of the influence of ploidy on evolution is among
such crops as wheat and cotton. Somy describes changes to the frequency of
particular chromosomes: for example, trisomy is three copies of a chromosome.
5.5
Population thinking
So far we have focused on how genes are inherited and how they or their
combinations can change. In order to understand evolutionary processes
(Chapter 4) we must shift our attention to looking at populations (we shall not
emphasize too much whether of genes. chromosomes, genomes, or organisms).
Population thinking is central to our understanding of models of evolution.
The Hardy-Weinberg theorem applies to frequencies of genes and genotypes
LRN the relative frequency of each gene
in a population of individuals, and TEAM
states that
36
Principles of genetics
remains in equilibrium from one generation to the next. For a single allele. if
the frequency of one form is p then that of the other (say 4 ) is I - p . The three
genotypes that exist with this allele have the population proportions of
pz +2py +y2 = I.
This equation does not apply when a mixture of four factors changes the relative
frequencies of genes in a population: mutation, selection, gene flow, and random
genetic drift (drift). Drift can be described as the effect of the sampling of a
population on its parents. Each generation can be thought of as a sample of its
parents' population. In that the current population is a sample of its parents,
we acknowledge that a statistical sampling error should be associated with gene
frequencies. The effect will be small in large populations because the relative
proportion of random changes will be a very small component of the large
numbers. However, drift in a small population will have a marked effect.
One factor which can counteract the effect of drift is differential migration
of individuals between populations which leads to gene flow. Several models of
gene flow exist. For example, migration which occurs at random among a group
of small populations is called the island model whereas in the stepping stone
model each population receives migrants only from neighboring populations.
Mutation, selection, and gene flow are deterministic factors so that if fitness.
mutation rate, and rate of gene flow are the same for a number of populations
that begin with same gene frequencies, they will attain the same equilibrium
composition. Drift is a stochastic process because the sampling effect on the
parent population is random.
Sewall Wright introduced the idea of an adaptive landscape to explain how
a population's allele freyuencies might evolve over time. The peaks on the
landscape represent genetic compositions of a population for which the mean
fitness is high and troughs are possible compositions where the mean fitness
is low. As gene frequencies change and mean fitness increases the population
moves uphill. Indeed, selection will operate to increase mean fitness so, on
a multipeaked landscape, selection may operate to move populations to local
maxima. On a fixed landscape drift and selection can act together so that
populations may move uphill (through selection) or downhill (through drift).
This means that the global maximum for the landscape could be reached. These
ideas are formally encapsulated in Wright's ( 1968-1 978) shifring htr1ciiic.e theory
of evolution. Further information on the relation of population genetics to
evolutionary theory can be studied further in the books by Wright (1968-1978),
Crow and Kimura (1970) and Maynard Smith (1989).
The change of gene frequencies coupled with changes in the genes
themselves can lead to the emergence of new species although the process
is far from simple and not fully understood (Futuyma, 1986, Maynard Smith
1993). The nature of the species concept or (for some) concepts which is
central to Darwinism is complicated and will not be discussed here (see e.g.
TEAMapply
LRN to promote speciation (Maynard
Futuyma 1986). Several mechanisms
Population thinking
37
38
Principles of genetics
References
Alberts B, Bray D, Lewis J, Raff M. Roberts K and Watson J D 1989 Molecitlrr Biology
of the Cell (New York: Garland)
Axelrod R 1984 The E\diitioti oJ Co-operation (Harmondsworth: Penguin)
Beauniont M A 1993 Evolution of optimal behaviour in networks of Boolean automata
J. Theor. Biol. 165 455-76
Changeux J-P and Dehaene S 1989 Neuronal models of cognitive functions Cognition
33 63- I09
Clarke B, Mittenthal J E and Senn M 1993 A model for the evolution of networks of
genes J . Theor. Biol. 165 269-89
Collins R 1994 Artificial evolution and the paradox of sex Compirting ktith Biolo,qicul
Mertrphor.s ed R C Paton (London: Chapman and Hall)
Crow J F and Kimura M 1970 An lritroduc~tiont o Popukrtion Genetics T h e o p (New(
York: Harper and Row)
Darden L 1991 Theory Chcrnge in Science (New York: Oxford University Press)
Darden L and Cain J A 1987 Selection type theories Phil. Sci. 56 106-29
Futuyma D J 1986 Eiwlittioncin Biology (MA: Sinauer)
Goodwin B C and Saunders P T (eds) 1989 Theoreticul Biology: Epigenetic. i i n d
Eivol ir t i o m Ord t.r from Conipl ex Systems ( Edinburgh : Edinburgh U n i ve rs i 1y
Press)
Hamilton W D, Axelrod A and Tanese R 1990 Sexual reproduction as an adaptation to
resist parasites Proca. Natl Acud. Sci. USA 87 3566-73
Hilario E and Gogarten J P 1993 Horizontal transfer of ATPase genes-the tree of life
becomes a net of life BioSj*stem.r31 1 1 1-9
Kauffman S A 1993 The Origiris of Order (New York: Oxford University Press)
Kimura, M I983 The Neirtrul Theot;v oj Molrc*itlar Eiwlirtioti (Cambridge: Cambridge
University Press)
Landman 0 E 199 1 The inheritance of acquired characteristics Anti. Re,: Genet. 25 1-20
Lewin B 1990 Genes IV (Oxford: Oxford University Press)
Lima de Faria A I988 E , d u t i o n Mithoirt Selection (Amsterdam: Elsevier)
Margulis L and Foster R (eds) 199I Symbiosis ( i s ci Soiirce ( f E ~ d u t i o n u nI t i n o i w t i o n :
Spwiiitioti ~ u i dMorphogeriesis (Cambridge, MA: MIT Press)
Manderick B 1994 The importance of selectionist systems for cognition Conipirting rcith
B i o l o g i d Metuphors ed R C Paton (London: Chapman and Hall)
Maynard Smith J 1989 E\dirrioncirj Genetics (Oxford: Oxford University Press)
____ I993 The Theor? ofEidirtioti Canto edn (Cambridge: Cambridge University Press)
Neidhart F C, Ingraham J L and Schaechter M 1990 Physiology of the Bucteritrl Cell
(Sunderland. MA: Sinauer)
Paton R C I994 Enhancing evolutionary computation using analogues of biological
mechanisms Eiwlittioncirj Conipirtirig (Lectitre Notes in Compicter Scienc*e 865) ed
T C Fogarty (Berlin: Springer) pp 51-64
Sigmund K 1993 Games ofLlfe (Oxford: Oxford University Press)
Sinnott E W, Dunn L C and Dobzhansky T 1958 Principles (J Genetics (New York:
McGraw-Hill)
Sober E 1984 The Nittirre oj Selection: E\diitionctt? Theorj in Philosophicd Foc-u.\
TEAM
LRN
Press)
(Chicago, IL: Unitwsity of Chicago
References
39
Sumida B and Hamilton W D 1994 Both Wrightian and 'parasite' peak shifts enhance
genetic algorithm performance in the travelling salesman problem Coniliictirig \r*ith
Biological Metriphors ed R C Paton (London: Chapman and Hall)
Van Valen L 1973 A new evolutionary law Eidutionurv Tiiror? 1 1-30
Wright S 1968-1978 E\dutioii und the Genetics of Popultrtioiis vols 1 4 (Chicago, IL:
Chicago University Press)
TEAM LRN
6.1
Introduction
No one will ever produce a completely accurate account of a set of past events
since. as someone once pointed out, writing history is as difficult as forecasting.
Thus we dare to begin our historical summary of evolutionary computation
rather arbitrarily at a stage as recent as the mid- 1950s.
At that time there was already evidence of the use of digital computer
models to better understand the natural process of evolution. One of the first
descriptions of the use of an evolutionary process for computer problem solving
appeared in the articles by Friedberg (1958) and Friedberg ef a1 (1959). This
represented some of the early work in machine learning and described the use
of an evolutionary algorithm for uutomcrtic proqrmimirig, i.e. the task of finding
a program that calculates a given input-output function. Other founders in the
field remember a paper of Fraser (1957) that influenced their early work, and
there may be many more such forerunners depending on whom one asks.
In the same time frame Brernermann presented some of the first attempts
to apply simulated evolution to numerical optimization problems involving both
linear and convex optimization as well as the solution of nonlinear simultaneous
equations (Bremermann 1962). Bremermann also developed some of the
early evolutionary algorithm (EA) theory, showing that the optimal mutation
probability for linearly separable problems should have the value of I/! in the
case of t bits encoding an individual (Bremermann et ul 1965).
Also during this period Box developed his evolutionary operation (EVOP)
ideas which involved an evolutionary technique for the design and analysis of
(industrial) experiments (Box 1957, Box and Draper 1969). Box's ideas were
never realized as a computer algorithm, although Spendley et a1 (1962) used
them as the basis for their so-called simplex design method. It is interesting to
note that the REVOP proposal (Satterthwaite 1959a, b) introducing randomness
into the EVOP operations was rejected at that time.
TEAM LRN
Evolutionary programming
41
As is the case with many ground-breaking efforts, these early studies were
met with considerable skepticism. However, by the mid-1960s the bases for
what we today identify as the three main forms of EA were clearly established.
The roots of evolutionary programming (EP) (Chapter 10) were laid by Lawrence
Fogel in San Diego, California (Fogel et a1 1966) and those of genetic algorithms
(GAS) (Chapter 8) were developed at the University of Michigan in Ann Arbor
by Holland (1967). On the other side of the Atlantic Ocean, evolution strategies
(ESs) (Chapter 9) were a joint development of a group of three students, Bienert,
Rechenberg, and Schwefel, in Berlin (Rechenberg 1965).
Over the next 25 years each of these branches developed quite independently
of each other, resulting in unique parallel histories which are described in more
detail in the following sections. However, in 1990 there was an organized effort
to provide a forum for interaction among the various EA research communities.
This took the form of an international workshop entitled Parallel Problem
Solling fkom Nature at Dortmund (Schwefel and Manner I99 1 ).
Since that event the interaction and cooperation among EA researchers from
around the world has continued to grow. In the subsequent years special efforts
were made by the organizers of ZCGA91 (Belew and Booker 1991), EP92
(Fogel and Atmar 1992), and PPSN92 (Manner and Manderick 1992) to provide
additional opportunities for interaction.
This increased interaction led to a consensus for the name of this new field,
e\vlutionarv computation (EC), and the establishment in 1993 of a journal by the
same name published by MIT Press. The increasing interest in EC was further
indicated by the IEEE World Congress on Computationcll Intelligence ( WCCI)
at Orlando, Florida, in June 1994 (Michalewicz et ml 1994), in which one of the
three simultaneous conferences was dedicated to EC along with conferences on
neural networks and fuzzy systems.
That brings us to the present in which the continued growth of the field is
reflected by the many EC events and related activities each year, and its growing
maturity reflected by the increasing number of books and articles about EC.
In order to keep this overview brief, we have deliberately suppressed many
of the details of the historical developments within each of the three main EC
streams. For the interested reader these details are presented in the following
sections.
42
Evolutionary programming
43
is required. The best machine generates this prediction, the new symbol is added
to the experienced environment, and the process is repeated. Fogel (1964) (and
Fogel et nl ( I 966)) used nonregressive evolution. To be retained, a machine
had to rank in the best half of the population. Saving lesser-adapted machines
was discussed as a possibility (Fogel et cil 1966, p 21) but not incorporated.
This general procedure was successfully applied to problems in prediction.
identification, and automatic control (Fogel et nl 1964, 1966, Fogel 1968) and
was extended to simulate coevolving populations by Fogel and Burgin (1969).
Additional experiments evolving finite-state machines for sequence prediction.
pattern recognition, and gaming can be found in the work of Lutter and
Huntsinger ( I969), Burgin ( 1969), Atmar ( 1976). Dearholt ( 1 9 7 6 ~ and
Takeuchi ( 1980).
In the mid-1980s the general EP procedure was extended to alternative
representations including ordered lists for the traveling salesman problem (Fogel
and Fogel 1986), and real-valued vectors for continuous function optimization
(Fogel and Fogel 1986). This led to other applications in route planning
(Fogel 1988, Fogel and Fogel 1988), optimal subset selection (Fogel 1989),
and training neural networks (Fogel et a1 1990), as well as comparisons to other
methods of simulated evolution (Fogel and Atmar 1990). Methods for extending
evolutionary search to a two-step process including evolution of the mutation
variance were offered by Fogel et nl (1991, 1992). Just as the proper choice of
step sizes is a crucial part of every numerical process, including optimization, the
internal adaptation of the mutation variance(s) is of utmost importance for the
algorithms efficiency. This process is called self-adaptation or autoadaptation
in the case of no explicit control mechanism, e.g. if the variances are part of
the individuals characteristics and underlie probabilistic variation in a similar
way as do the ordinary decision variables.
In the early 1990s efforts were made to organize annual conferences on EP,
these leading to the first conference in 1992 (Fogel and Atmar 1992). This
conference offered a variety of optimization applications of EP in robotics
(McDonnell et a1 1992, Andersen er a1 1992). path planning (Larsen and
Herman 1992, Page et a1 1992), neural network design and training (Sebald
and Fogel 1992, Porto 1992, McDonnell 1992), automatic control (Sebald et nl
19921, and other fields.
First contacts were made between the EP and ES communities just
before this conference, and the similar but independent paths that these two
approaches had taken to simulating the process of evolution were clearly
apparent. Members of the ES community have participated in all successive
EP conferences (Back et a1 1993, Sprave 1994, Back and Schiitz 1995, Fogel et
(11 1996). There is less similarity between EP and GAS, as the latter emphasize
simulating specific mechanisms that apply to natural genetic systems whereas
EP emphasizes the behavioral, rather than genetic, relationships between parents
and their offspring. Members of the GA and GP communities have, however,
LRN
also been invited to participate inTEAM
the annual
conferences, making for truly
44
interdiwiplinary interaction (see e.g. Altenberg 1994, Land and Belew 1995,
Koza and Andre 1996).
Since the early 1990s. efforts i n EP have diversified in many directions.
Application\ in training neural networks have received considerable attention
( w e e.g. English 1994, Angeline st d 1994, NlcDonnell and Waagen 1994,
Porto et 111 199.5). while relatively less attention ha\ been devoted to evolving
furry \ystems (Haffner and Sebald 1993, Kim and Jeon 1996). Image processing
applications can be found in the articles by Bhattacharjya and Roysam (1994).
Brotherton et trl ( 1994). Rizki et til (1995), and others. Recent efforts to use
EP in medicine hace been offered by Fogel et cil (1995) and Gehlhaar et trl
( 1995 1.
Efforts \tudying and comparing methods of self-adaptation can be
found in the articles by Saralanan et crl ( 1995). Angeline er al ( 1996). and
other\. Mathematical analyse\ of EP have been summarized by Fogel (1995).
To offer a summary, the initial efforts of L J Fogel indicate some of the
early attempts to ( i ) use simulated cvolution to perform prediction, ( i i ) include
ariable-length encodings, (iii) use representation\ that take the form of a
\equence of instructions. ( i v ) incorporate a population of candidate solutions, and
( b ) coevolve evolutionary programs. Moreover, Fogel ( 1963, 1964) and Fogel
et trl (1966) offered the early recognition that natural evolution and the human
endeavor of the scientific method are essentially \imilar processes, a notion
recently echoed by Cell-Mann ( 1994). The initial prescription3 for operating
on tinite-state machines have been extended to arbitrary representations.
mutation operators, and selection methods, and techniques for self-adapting the
t ' b olutionary search hake been proposed and implemented. The population sire
need not be kept constant and there can be a variable number of offjpring
per parent, much like the ( p A ) methods (Section 25.4) offered in ESs. In
contra\t to these methods, selection is often made probabilistic in EP, giving
les\er-\coring solutions \ome probability of surviving as parents into the next
generation. In contrast to GAS, no effort is made in EP to support (some \ay
maximim) schema processing. nor is the use of random variation comtrained
to emphasize \pecitic mechanisms of genetic transfer, perhaps providing greater
\ ersatility to tackle specific problem domains that are unsuitable for genetic
operator\ wch a\ cros\over.
Genetic algorithms
45
forms of adaptation as they appear in natural systems and our ability to design
robust adaptive artifacts.
In Hollands view the key feature of robust natural adaptive systems
was the successful use of competition and innovation to provide the ability
to dynamically respond to unanticipated events and changing environments.
Simple models of biological evolution were seen to capture these ideas nicely via
notions of survival of the fittest and the continuous production of new offspring.
This theme of using evolutionary models both to understand natural adaptive
systems and to design robust adaptive artifacts gave Hollands work a somewhat
different focus than those of other contemporary groups that were exploring the
use of evolutionary models in the design of efficient experimental optimization
techniques (Rechenberg 1965) or for the evolution of intelligent agents (Fogel
et a1 1966), as reported in the previous section.
By the mid-1960s Hollands ideas began to take on various computational
forms as reflected by the PhD students working with Holland. From the outset
these systems had a distinct genetic flavor to them in the sense that the
objects to be evolved over time were represented internally as genomes and the
mechanisms of reproduction and inheritance were simple abstractions of familiar
population genetics operators such as mutation, crossover, and inversion.
Bagleys thesis (Bagley 1967) involved tuning sets of weights used in the
evaluation functions of game-playing programs, and represents some of the
earliest experimental work in the use of diploid representations. the role of
inversion, and selection mechanisms. By contrast Rosenbergs thesis (Rosenberg
1967) has a very distinct flavor of simulating the evolution of a simple
biochemical system in which single-celled organisms capable of producing
enzymes were represented in diploid fashion and were evolved over time to
produce appropriate chemical concentrations. Of interest here is some of the
earliest experimentation with adaptive crossover operators.
Cavicchios thesis (Cavicchio 1970) focused on viewing these ideas as a form
of adaptive search, and tested them experimentally on difficult search problems
involving subroutine selection and pattern recognition. In his work we see
some of the early studies on elitist (section 28.4) forms of selection and ideas
for adapting the rates of crossover and mutation. Hollstiens thesis (Hollstien
1971) took the first detailed look at alternate selection and mating schemes.
Using a test suite of two-dimensional fitness landscapes, he experimented with
a variety of breeding strategies drawn from techniques used by animal breeders.
Also of interest here is Hollstiens use of binary string encoding of the genome
and early observations about the virtues of Gray codings.
In parallel with these experimental studies, Holland continued to work on
a general theory of adaptive systems (Holland 1967). During this period he
developed his now famous schema analysis of adaptive systems, relating it to
the optimal allocation of trials using k-armed bandit models (Holland 1969).
He used these ideas to develop a more theoretical analysis of his reproductive
TEAM
LRN Holland then pulled all of these
plans (simple GAS) (Holland 1971,
1973).
46
ideas together in his pivotal book AdnptLition in h'oturul mid Artijcicil Systems
(Holland 1975).
Of interest was the fact that many of the desirable properties of these
algorithms being identified by Holland theoretically were frequently not
observed experimentally. It was not difficult to identify the reasons for this.
Hampered by a lack of computational resources and analysis tools, most of
the early experimental studies involved a relatively small number of runs using
small population sizes (generally less than 20). It became increasingly clear
that many of the observed deviations from expected behavior could be traced
t o the well-known phenomenon in population genetics of genetic d r i f , the loss
of genetic diversity due to the stochastic aspects of selection, reproduction, and
the like in small populations.
By the early 1970s there was considerable interest in understanding better
the behavior of implementable GAS. In particular, it was clear that choices
of population size, representation issues, the choice of operators and operator
rates all had significant effects of the observed behavior of GAS. Frantz's thesis
(Frantz 1972) reflected this new focus by studying in detail the roles of crossover
and inversion in populations of size 100. Of interest here is some of the earliest
experimental work on mu1ti poin t crossover operators.
De Jong's thesis (De Jong 1975) broaded this line of study by analyzing
both theoretically and experimentally the interacting effects of population size.
crossover, and mutation on the behavior of a family of GAS being used to
optimize a fixed test suite of functions. Out of this study came a strong sense that
even these simple GAS had significant potential for solving difficult optimization
pro b 1ems.
The mid-1970s also represented a branching out of the family tree of GAS
as other universities and research laboratories established research activities in
this area. This happened slowly at first since initial attempts to spread the word
about the progress being made in GAS were met with fairly negative perceptions
from the artificial intelligence (AI) community as a result of early overhyped
work in areas such as self-organizing systems and perceptrons.
Undaunted, groups from several universities including the University of
Michigan, the University of Pittsburgh, and the University of Alberta organized
an Aikiptii'e Systerns Worksliop in the summer of 1976 in Ann Arbor, Michigan.
About 20 people attended and agreed to meet again the following summer. This
pattern repeated itself for several years, but by 1979 the organizers felt the
need to broaden the scope and make things a little more formal. Holland, De
Jong, and Sampson obtained NSF funding for Ail 1)zterdisi~iplincir~
Workshop iiz
,4ticiptii*eS~stems,which was held at the University of Michigan in the summer
of I98 1 (Sampson I98 I ).
By this time there were several established research groups working on GAS.
At the University of Michigan, Bethke, Goldberg, and Booker were continuing
to develop GAS and explore Holland's classifier systems (Chapter 12) as part
TEAM
LRN
o f their PhD research (Bethke 1981,
Booker
1982, Goldberg 1983). At the
Genetic algorithms
47
48
and a growing list of journal papers. New paradigms such as messy GAS
(Goldberg et a1 1991) and genetic programming (Chapter 1 1 ) (Koza 1992)
were being developed. The interactions with other EC communities resulted
in considerable crossbreeding of ideas and many new hybrid EAs. New GA
applications continue to be developed, spanning a wide range of problem areas
from engineering design problems to operations research problems to automatic
programming.
Evolution strategies
49
so
References
51
Rudolph 1995, Back and Schwefel 1995, Schwefel and Bick 1995), which on
the one hand define the actual standard ES algorithms and on the other hand
present some recent theoretical results.
References
Altenberg L I994 Emergent phenomena in genetic programming Proc. 3rd Aiinic. Conf
oii E\dirtionLiny Progmtnniing (San Diego, CA, 1994) ed A V Scbald and L J Fogel
(Singapore: World Scientific) pp 2 3 3 4 1
Andersen B, McDonnell J and Page W 1992 Configuration optimization of mobile
manipulators with equality constraints using evolutionary programming Proc. I \t
Atin. Conf oii Evolutionuq~Progranzming (Lci J o l l a , CA, 1992) ed D B Fogel and
W Atmar (La Jolla, CA: Evolutionary Programming Society) pp 71-9
Angeline P J, Fogel D B and Fogel L J 1996 A compariwn of self-adaptation methods
for finite \(ate machines in a dynamic environment Eidirtionut?~ProgruimiinCq
V-Proc. 5th Ann. Cot$ on E\dutinnai? Progrurnriiirig ( 1996) ed L J Fogel,
P J Angeline and T Back (Cambridge, MA: MIT Pre\s)
Angeline P J, Saunders G M and Pollack J B 1994 An evolutionary algorithm that
constructs recurrent neural networks IEEE Trrrns. Neirrul Nont*orhs " - 5 54-65
ofthe E,wlution c?flntelligc.nc.t~
u r i d I t s Possihlt~Reuli:trtiori
Atmar J W 1976 S~iec~irlcitioti
in Muclzine Form ScD Thesis, New Mexico State University
Back T I996 E\vlutionun Algorithriis iiz Theor? und Pructice (New York: Oxford
University Pre\s)
Back T, Hoffmeister F and Schwefel H-P 1991 A wrvcy of evolution \trategies Pmc.
4th Int. Conf. on Genetic Algorithitis (Sun Diego, CA, 1991I ed R K Belew and
L B Booker (San Mateo, CA: Morgan Kaufmann) pp 2-9
-1
992 Applicutions of E~dutioncrn?
Algorithnis Technical Report o f the Uni\wsity
of Dortmund Department of Computer Science Sy\tem\ Analy5is Rcwarch Group
sY s-2/92
Back T, Rudolph G and Schwefel H-P 1993 Evolutionary programming and evolution
strategies: similarities and differences Proc. 2nd Anti. Conf. on E\wlutionrrr~
Progrurnniing (San Diego, CA, 1993) ed D B Fogel and W Atmar (La Jolla. CA:
Evolutionary Programming Society) pp 1 1-22
Back T and Schutz M 1995 Evolution strategies for mixed-integer optimization of
optical multilayer systems Evolutionan Progrumrning IV-Pro(.. 4th Ann. Conf 0 1 1
E\dutionur? Programming (Sun Diego, CA, 1995) ed J R McDonnell, R G Reynolds
and D B Fogel (Cambridge, MA: MIT Press) pp 33-51
Back T and Schwefel H-P 1995 Evolution strategies I: variant5 and their computational
implementation Genetic A Igorithms in Engineering und Coitiptrr Scirticr, Proc. I J t
Short Coiirse EUROGEN-95 ed G Winter, J Pkriaux, M Galhn and P Cuesta (New
York: Wiley) pp I 1 1-26
Bagley J D 1967 The Behavior of Adup five Systems \r%ic.h Employ Griwtic mid
Correlution Algorithms PhD Thesis, University of Michigan
Belew R K and Booker L B (eds) 199 1 Proc. 4th liit. Cot$ on Genrtic AI~qorithiti.s(Strii
Diego, CA, 1991) (San Mateo, CA: Morgan Kaufmann)
Bethke A D 198 1 Genetic Algorithriis us Futicrion Optimi,-er\ PhD Thesis, Univer4ty of
TEAM LRN
Michigan
52
References
53
Fogel D B and Atmar J W 1990 Comparing genetic operators with Gaussian mutations
in simulated evolutionary processing using linear systems Biol. Cyhernet. 63 1 1 1 - 4
-(eds)
1992 Proc. 1st A m . Con8 on Elvlutiortan Progrunttning (LA Jolla, CA, 1992)
(La Jolla, CA: Evolutionary Programming Society)
Fogel D B and Fogel L J 1988 Route optimization through evolutionary programming
Proc. 22nd Asilornur Conf on Signals, Systems und Computers (Pocijk Grmv. CA )
pp 679-80
Fogel D B, Fogel L J and Atmar J W 1991 Meta-evolutionary programming Pro(.. 25th
Asilontar Conf: on Signals, Systems and Compitters (Pncific Gro\te, CA) ed R R Chen
pp 540-5
Fogel D B, Fogel L J, Atmar J W and Fogel G B 1992 Hierarchic methods of evolutionary
programming Pro(.. I st Anti. Con8 on E\dutionuty Progruntntin~q(Lu Jollu, CA,
1992) ed D B Fogel and W Atmar (La Jolla, CA: Evolutionary Programming
Society) pp 175-82
Fogel D B, Fogel L J and Port0 V W 1990 Evolving neural networks Biol. Cyhurnet. 63
487-93
Fogel D B, Wasson E C and Boughton E M 1995 Evolving neural networks for detecting
breast cancer Cancer Lett. 96 49-53
Fogel L J 1962 Autonomous automata Industrial Res. 4 14-9
-1963 Biotechnology: Corzcepts and Applications (Englewood Cliffs, N J : PrenticeHall)
-I964 On the 0rgcirti:cition of Intellect PhD Thesis, University of California at Los
Angeles
-1
968 Extending communication and control through simulated evolution
Bioengineering-urt Engineering Veu. Proc. Sytnp. on En~qineeringSi~qniJic~unc~u
of'
the Biological Sciences ed G Bugliarello (San Francisco, CA: San Francisco Press)
pp 2 8 6 3 0 4
Fogel L J, Angeline P J and Back T (eds) 1996 Evolutionun Progrumming V-Proc. 5th
Atin. Conf on Evolutionan Programming (1996) (Cambridge, MA: MIT Press)
Fogel L J and Burgin G H 1969 Competitive Goal-seeking throirgh E\~olutionctr?
Progrumming Air Force Cambridge Research Laboratories Final Report Corltract
AF 19(628)-5927
Fogel L J and Fogel D B 1986 Artijicial Iiztelligence through E\dutioitun Progrutnniin~q
U S Army Research Institute Final Report Contract P0-9-X56- 1 102C-1
Fogel L J, Owens A J and Walsh M J 1964 On the evolution of artificial intelligence
Proc. 5th Natl Syntp. on Human Factors in Electronic*.s(San Diego, CA: IEEE)
-1
965 Artificial intelligence through a simulation of evolution Biophysics und
Cybernetic Sjstems ed A Callahan, M Maxfield and L J Fogel (Washington, DC:
Spartan) pp 13 1-56
-1966 Artijicial Intelligence through Simulated E\dution (New York: Wiley)
FrantL D R 1972 Non-lirtearities in Genetic Aduptilre Setrrch PhD Thesis, University of
Michigan
Fraser A S 1957 Simulation of genetic systems by automatic digital computers Aust. J.
Biol. Sci. 10 484-99
Friedberg R M 1958 A learning machine: part I IBM J. 2 2-1 3
Friedberg R M, Dunham B and North J H 1959 A learning machine: part I1 IBM .I.
3
TEAM LRN
282-7
54
Furst H, Muller P H and Nollau V 1968 Eine stochastische Methode zur Ermittlung
der Maximalstelle einer Funktion von mehreren Veranderlichen mit experimentell
ermittelbaren Funktionswerten und ihre Anwendung bei chemischen Prozcssen
c i ~ e t t t . - w h 20
. w - s
Gehlhaar et cil I99SGehlhaar D K et cil 19% Molecular recognition of the inhibitor
AG- 1343 by HIV- I protease: conformationally flexible docking by evolutionary
programming Clienz. Biol. 2 3 17-24
Cell-Mann M 1994 The Q i ~ i r kcintl the J q i i a r (New York: Freeman)
Goldberg D E 1983 Coinpiiter-Aided Gris Pipeliiie Opemtioti icsing Genetic A1goritlitn.s
ciiitl Ride Lecirtiirzg PhD Thesis, University of Michigan
1989 Genetic Algorithnis iri Seiirch, Optimiz~itiotir i d Mnchine Leiirtiirig (Reading,
MA: Addison- Wesley )
Goldberg D E. Deb K and Korb B 1991 Don't worry, be messy Proc-. 4th Int. Corzf. 011
Gerirtic Algorithr?i.s ( S m Diego, CA, I W l ) ed R K Belew and L B Booker (San
Mateo, CA: Morgan Kaufmann) pp 24-30
Grefenstette J J (ed) 1985 Pro(-. 1st Int. Conf: 011 Genetic Algorithnis mid Their
Applic*doris (Pittshitrgti, PA, 198.5) ( Hillsdale, NJ: Erlbaum)
__ I987 Proc. 2nd lnt. Cot$ o ~ G
i e w t i c Algorithms mcl Their Applicutiom (Cmihrirl~qr,
MA. 1987) (Hillsdale, NJ: Erlbaum)
Haffner S B and Sebald A V 1993 Computer-aided design of fuzzy HVAC
controllers using evolutionary programming Proc. 2nd Ann. Cot$ o i l E \ ~ ) / i i t i o i i ~ i r ~ ~
Pr-ogrcrtnmhg (Sciti Diego, CA, 1993) ed D B Fogel and W Atmar (La Jolla, CA:
Evolutionary Programming Society) pp 98- 107
Heydt G T 1970 Directed R ~ i t i d o mSetirdi PhD Thesis, Purdue University
Hoftmeister F and Schwefcl H-P I990 A taxonomy of parallel evolutionary algorithms
Ptircellu '90. Proc. 5th Int. Workshop on Prircillel Processing b y Cellitlcir A utonirrtci
w i t 1 Arrciys vol 2 , cd G Wolf, T Legendi and U Schendcl (Berlin: Academic)
pp 97-107
Holland J H 1962 Outline for a logical theory of adaptive systems J. ACM 9 297-313
__ 1967 Nonlinear environments permitting efficient adaptation CotnpicttJr md
Infbrnztition Scietic~~s
I1 (New York: Academic)
__ I969 Adaptive plans optimal for payoff-only environments Proc. 2nd Hm-trii Int.
Cot$ oti System Sciences pp 9 17-20
197 1 Processing and processors for schemata A s s o c i d \ v it~fi)rmitioiiprocessitiq ed
E L Jacks (New York: Elsevier) pp 127-46
-1973
Genetic algorithms and the optimal allocation of trials SIAM J. Conzput. 2
88- 1 OS
I975 Atlqitiition in Ntitirnil md Artijickil Systetns ( Ann Arbor, MI: University of
Michigan Press)
Hollsticn R B I97 I Ar-tijkitil Genrtic Ackiptcirioti in Coniiiicter Coritrol Systenzs PhD
Thesis, University of Michigan
Kim J-H and Jeon J-Y 1996 Evolutionary programming-based high-precision controller
design E\diitiottcit:\' Progmmniitig V-Pro(.. 5th Anti. Cord; 011 E\~olirtioticir\~
Progrcrmmitig (1996) ed L J FogeI, P J Angeline and T Back (Cambridgc, MA:
MIT Press)
Klockgcther J and Schwefel H-P 1970 Two-phase nozzle and hollow core jet experiments
Proc. I ltli Sytrip. 0 1 1 Engitieeririg Aspects c?f'~.kc~~,zetotiy~lr~)dyti~i~~iic.s
ed D G Elliott
TEAM
LRN
(Pasadena. CA: California Institute
of Technology)
pp 14 1-8
~
References
55
56
References
57
58
TEAM LRN
7
Introduction to evolutionary algorithms
Thornas Back
7.1
Since they are gleaned from the model of organic evolution, all basic instances
of evolutionary algorithms share a number of common properties, which are
mentioned here to characterize the prototype of a general evolutionary algorithm:
(0
59
60
61
Input:
Output:
I
2
3
4
5
6
7
8
9
@,
U * ,the
to;
P ( t ) t initialize(p);
F ( t ) +- evaluate( P ( t ) , p ) ;
while ( i ( P ( t ) ,0 , )# true) do
P ( t ) t recombine(P(t), 0,);
P(r) t mutate(P(t), ( E n l ) ;
F ( t ) t evaluate(P(t), A);
P ( t 1 ) +- select(P(t), F ( t ) ,p , (-1,);
t t + l ;
od
After initialization of t (line 1 ) and the population P ( t ) of size p (line 2) as
well as its fitness evaluation (line 3), the while-loop is entered. The termination
criterion i might depend on a variety of parameters, which are summarized
here by the argument 0,. Similarly, recombination (line 5 ) , mutation (line
6), and selection (line 8) depend on a number of algorithm-specific additional
parameters. While P ( t ) consists of p individuals, P ( t ) and P ( t ) are assumed
to be of size K and A, respectively. Of course, h = K = 1-l is allowed and
is the default case in genetic algorithms. The setting K = p is also often
used in evolutionary programming (without recombination), but it depends on
the application and the situation is quickly changing. Either recombination
or mutation might be absent from the main loop. such that K = p (absence
of recombination) or K = h (absence of mutation) is required in these cases.
The selection operator selects p individuals from P ( t ) according to the fitness
values F ( t ) ,t is incremented (line 9),
andLRN
the body of the main loop is repeated.
TEAM
62
References
Back T I996 Eivlictionuq~Algorithms in Theory Litid Prcrc-ticv (New York: Oxford
University Press)
Back T and Schwefel H-P 1993 An overview of evolutionary algorithms for parameter
optimization Eidi.rtionaq Coinputdon 1( 1 ) 1-23
Fogcl D B I992 E\*ol\+ig Art$ciril Intelligent-e PhD Thesis, University of California,
San Diego
1995 E iwlu tiotiti n v Comyutii tioii : h ure1 ei Nr \t Philosophy OJ Meicli in e In tell ig ence
(Piscataway, NJ: IEEE)
Fogel L J 1962 Autonomous automata Industr. Res. 4 14-9
Fogel L J, Owens A J and Walsh M J 1966 ArtiJicWl 1ntelligetic.e throicgh Sitniilcit(d
Eiwliition (New York: Wiley)
Holland J H 1962 Outline for a logical theory of adaptive systems J. ACM 3 297-314
__ 1975 Atkiptcition in Natiiml m d Articiul
System (Ann Arbor, MI: University of
Michigan Press)
Rechenberg I I965 Cybernetic solution path of an experimental problem Lihrtiry
Trcinsl~itionN o I I22 Royal Aircraft Establishment, Farnborough, U K
___ 1 973 Eioliitioiisstrtitegie: Optitnierung techiischer Systerne n c i c 4 Priii:ipirn tier
hiolog isc -hen E\ditt ion ( Stuttgart : Fromman n- Ho 1I,boog )
Sch we fel H-P 1965 K.vhertietisc.heEi)olution a l s Strcctegie der experimentellrti Forschng
iri der Striirizittig.stec.htzik Diplomarbeit, Technische Universitiit, Berlin
__ I 97 7 Nittnerist he Optitnierirtig \ w i Computer-Moclellen mittuls der E~olictioti.s.str~itegie Interdisciplinary Systems Research, vol 26 (Basel: Birkhiiuser)
~
Further reading
The introductory section to evolutionary algorithms certainly provides the right
place to mention the most important books on evolutionary computation and its
subdisciplines. The following list is not intended to be complete, but only to
guide the reader to the literature.
I. Back T 1996 E\dictiotitin Algorithms in Tlieot? and Pructic-r (New York: Oxford
University Press)
A presentation and comparison of evolution strategies, evolutionary programming.
and genetic algorithms with respect to their behavior as parameter optimization
methods. Furthermore, the role of mutation and selection in genetic algorithms is
discussed in detail, arguing that mutation is much more useful than usually claimed
TEAM LRN
in connection with genetic algorithms.
63
Further reading
Programs
This collection of articles summarizes the state of the art in genetic programming,
emphasizing other than LISP-based approaches to genetic programming.
8. Koza J R 1992 Genetic Programming: On the Progrumming of Computers h j Meurrs
of Nutural Selection (Cambridge, MA: MIT Press)
9. Koza J R 1994 Genetic Programming I1 (Cambridge. MA: MIT Press)
The basic books for genetic programming using LISP programs, demonstrating
the feasibility of the method by presenting a variety of application examples from
diverse fields.
TEAM LRN
Genetic algorithms
Larry J Eshelman
8.1 Introduction
Genetic algorithms (GAS) are a class of evolutionary algorithms first proposed
and analyzed by John Holland ( 1975). There are three features which distinguish
GAS, as first proposed by Holland, from other evolutionary algorithms: ( i )
the representation used-bitstrings (Chapter 15); (ii) the method of selectionproportional selection (Chapter 23) ; and (iii) the primary method of producing
variations-crossover (Chapter 33). Of these three features, however, it
is the emphasis placed on crossover which makes GAS distinctive. Many
subsequent GA implementations have adopted alternative methods of selection,
and many have abandoned bitstring representations for other representations
more amenable to the problems being tackled. Although many alternative
methods of crossover have been proposed, in almost every case these variants
are inspired by the spirit which underlies Hollands original analysis of GA
behavior in terms of the processing of schemata or building blocks. It should be
pointed out, however, that the evolution strategy paradigm (Chapter 9) has added
crossover to its repertoire, so that the distinction between classes of evolutionary
algorithms has become blurred (Bick et a1 1991).
We shall begin by outlining what might be called the canonical GA, similar
to that described and analyzed by Holland (1975) and Goldberg (1987). We
shall introduce a framework for describing GAS which is richer than needed
but which is convenient for describing some variations with regard to the
method of selection. First we shall introduce some terminology. The individual
structures are often referred to as chromosomes. They are the genotypes that
are manipulated by the GA. The evaluation routine decodes these structures
into some phenotypical structure and assigns a fitness value. Typically, but not
necessarily, the chromosomes are bitstrings. The value at each locus on the
bitstring is referred to as an allele. Sometimes the individuals loci are also
called genes. At other times genes are combinations of alleles that have some
phenotypical meaning, such as parameters.
TEAM LRN
65
After the new offspring have been created via the genetic operators the two
populations of parents and children must be merged to create a new population.
Since most GAS maintain a fixed-sized population M , this means that a total
of M individuals need to be selected from the parent and child populations to
create a new population. One possibility is to use all the children generated
(assuming that the number is not greater than M ) and randomly select (without
any bias) individuals from the old population to bring the new population up
to size M . If only one or two new offspring are produced, this in effect means
randomly replacing one or two individuals in the old population with the new
offspring. (This is what Hollands original proposal did.) On the other hand, if
TEAM to
LRN
the number of offspring created is equal
M , then the old parent population is
66
Genetic algorithms
67
of the M members of the child population are chosen. Depending upon the
implementation, the selection of the child to be replaced by the best indiLidual
from the parent population may or may not be biased.
A number of CA variations make use of biased replacement selection.
Whitleys GENITOR, for example, creates one child each cycle, selecting the
parents using ranked selection, and then replacing the worst member of the
population with the new child (Whitley 1989). Syswerdas steady-state CA
creates two children each cycle, selecting parents using ranked selection, and
then stochastically choosing two individuals to be replaced. with a bias towards
the worst individuals in the parent population (Syswerda 1989). Eshelmans
CHC uses unbiased reproductive selection by randomly pairing all the members
of the parent population, and then replacing the worst individuals of the parent
population with the better individuals of the child population. (In effect. the
offspring and parent populations are merged and the best M (population size)
individuals are chosen.) Since the new offspring are only chosen by CHC if
they are better than the members of the parent population. the selection of both
the offspring and parent populations is biased (Eshelman I99 1 ).
These methods of replacement selection, and especially that of CHC,
resemble the (1-1+A) ES method of selection (Section 25.4) sometimes originally
used by evolution strategies (ESs) (Back et (11 1991). From j1 parents h
offspring are produced; the p parents and A. offspring are merged; and the
best j i individuals are chosen to form the new parent population. The other
ES selection method, ( p , A ) ES (Section 25.4), places all the bias in the child
selection stage. In this case, j i parents produce h offspring ( E , > j i ) , and the best
p offspring are chosen to replace the parent population. Muhlenbeins breeder
CA also uses this selection mechanism (Muhlenbein and Schlierkamp-Voosen
1993).
Often a distinction is made between generational and stmdj~-state GAS
(Section 28.3). Unfortunately, this distinction tends to merge two properties that
are quite independent: whether the replacement strategy of the GA is biased
or not and whether the C A produces one (or two) versus many (usually M )
offspring each cycle. Syswerdas steady-state CA, like Whitleys GENITOR,
allows only one mating per cycle and uses a biased replacement selection,
but there are also GAS that combine multiple matings per cycle with biased
replacement selection (CHC) as well as a whole class of ESs ( ( p + E , ) ES).
Furthermore, the GA described by Holland (1975) combined a single mating per
cycle and unbiased replacement selection. Of these two features, it would seem
that the most significant is the replacement strategy. De Jong and Sarma (1993)
found that the main difference between GAS allowing many matings versus few
matings per cycle is that the latter have a higher variance in performance.
The choice between a biased and an unbiased replacement strategy, on the
other hand, is a major determinant of C A behavior. First, if biased replacement
is used in combination with biased reproduction, then the problem of premature
TEAM LRN(Of course this will depend upon
convergence is likely to be compounded.
68
Genetic algorithms
other factors, such as the size of the population, whether ranked selection is
used, and, if so, the setting of the selection bias parameter.) Second, the obvious
shortcoming of unbiased replacement selection can turn out to be a strength. On
the negative side, replacing the parents by the children, with no mechanism for
keeping those parents that are better than any of the children, risks losing,
perhaps forever, very good individuals. On the other hand, replacing the
parents by the children can allow the algorithm to wander, and it may be
able to wander out of a local minimum that would trap a GA relying upon
biased replacement selection. Which is the better strategy cannot be answered
except in the context of the other mechanisms of the algorithm (as well as the
nature of the problem being solved). Both Syswerdas steady-state GA and
Whitleys GENITOR combine a biased replacement strategy with a mechanism
for eliminating children which are duplicates of any member in the parent
population. CHC uses unbiased reproductive selection, relying solely upon
biased replacement selection as its only source of selection pressure, and uses
several mechanisms for maintaining diversity (not mating similar individuals and
seeded restarts), which allow it to take advantage of the preserving properties
of a deterministic replacement strategy without suffering too severely from its
shortcomings.
69
these features as building blocks scattered throughout the population and tries to
recombine them into better individuals via crossover. Sometimes crossover will
combine the worst features from the two parents, in which case these children
will not survive for long. But sometimes it will recombine the best features from
two good individuals, creating even better individuals, provided these features
are compatible.
Suppose that the representation is the classical bitstring representation:
individual solutions in our population are represented by binary strings of zeros
and ones of length L . A GA creates new individuals via crossover by choosing
two strings from the parent population, lining them up, and then creating two
new individuals by swapping the bits at random between the strings. (In some
GAS only one individual is created and evaluated, but the procedure is essentially
the same.) Holland originally proposed that the swapping be done in segments,
not bit by bit. In particular, he proposed that a single locus be chosen at random
and all bits after that point be swapped. This is known as one-point crossover.
Another common form of crossover is two-point crossover which involves
choosing two points at random and swapping the corresponding segments from
the two parents defined by the two points. There are of course many possible
variants. The best known alternative to one- and two-point crossover is i i i ~ l f o r i ? ~
crosso\er.Uniform crossover randomly swaps individual bits between the two
parents (i.e. exchanges between the parents the values at loci chosen at random).
Following Holland, GA behavior is typically analyzed in terms of schemata.
Given a space of structures represented by bitstrings of length L . schemata
represent partitions of the search space. If the bitstrings of length L are
interpreted as vectors in a L-dimensional hypercube, then schemata are
hyperplanes of the space. A schema can be represented by a string of L symbols
from the set 0, I , # where # is a wildcard matching either 0 or 1. Each string
of length L may be considered a sample from the partition defined by a schema
if it matches the schema at each of the defined positions (i.e. the non-# loci).
For example, the string 01 1001 instantiates the schema 01##0#. Each string, in
fact, instantiates 2 L schemata.
Two important schema properties are order and defining length. The order of
a schema is the number of defined loci (i.e. the number of non-# symbols). For
example the schema #01##1### is an order 3 schema. The defining length is
the distance between the loci of the first and last defined positions. The defining
length of the above schema is four since the loci of the first and last defined
positions are 2 and 6.
From the hyperplane analysis point of view, a GA can be interpreted as
focusing its search via crossover upon those hyperplane partition elements that
have on average produced the best-performing individuals. Over time the search
becomes more and more focused as the population converges since the degree of
variation used to produce new offspring is constrained by the remaining variation
in the population. This is because crossover has the property that Radcliffe refers
TEAM LRN
to as respect-if two parents are instances
of the same schema, the child will
70
Genetic algorithms
71
72
Genetic algorithms
as one- or two-point crossover and assuming that the important building blocks
are of short defining length. Unfortunately, for the types of problem to which
GAS are supposedly ideally suited-those that are highly complex with no
tractable analytical solution-there is no a priori reason to assume that the
problem will, or even can, be represented so that important building blocks will
be those with short defining length. To handle this problem Holland proposed an
inversion operator that could reorder the loci on the string, and thus be capable
of finding a representation that had building blocks with short defining lengths.
The inversion operator, however, has not proven sufficiently effective in practice
at recoding strings on the fly. To overcome this linkage problem, Goldberg has
proposed what he calls messy GAS, but, before discussing messy GAS, it will
be helpful to describe a class of problems that illustrate these linkage issues:
deceptive problems.
Deception is a notion introduced by Goldberg (1987). Consider two
incompatible schemata, A and B . A problem is deceptive if the average fitness of
A is greater than B even though B includes a string that has a greater fitness than
any member of A. In practice this means that the lower-order building blocks
lead the GA away from the global optimum. For example, consider a problem
consisting of five-bit segments for which the fitness of each is determined as
follows (Liepins and Vose 1991). For each one the segment receives a point,
and thus five points for all ones, but for all Term it receives a value greater
than five. For problems where the value of the optimum is between five and
eight the problem is fully deceptive (i.e. all relevant lower-order hyperplanes
lead toward the deceptive attractor). The total fitness is the sum of the fitness
of the segments.
It should be noted that it is probably a mistake to place too much emphasis on
the formal definition of deception (Grefenstette 1993). What is really important
is the concept of being misled by the lower-order building blocks. Whereas
the formal definition of deception stresses the average fitness of the hyperplanes
taken over the entire search space, selection only takes into account the observed
average fitness of hyperplanes (those in the actual population). The interesting
set of problems is those that are misleading in that manipulation of the lowerorder building blocks is likely to lead the search away from the middle-level
building blocks that constitute the optimum solution, whether these middle-level
building blocks are deceptive in the formal sense or not. In the above class of
functions, even when the value of the optimum is greater than eight (and so
not fully deceptive), but still not very large, e.g. ten, the problem is solvable
by a GA using segment-based crossover, very difficult for a GA using bitwise
uniform crossover, and all but impossible for a poolwise-based algorithm like
BSC.
As long as the deceptive problem is represented so that the loci of the
positions defining the building blocks are close together on the string. i t meets
Hollands original assumption that the important building blocks are of short
TEAM
defining length. The GA will be able
to LRN
exploit this information using one- or
73
14
Genetic algorithms
Representation
75
building blocks can provide an added value over CCV. It still is an open
question, however, as to how representative deceptive problems are of the types
of real-world problem that GAS might encounter. No doubt, many difficult
real-world problems have deceptive or misleading elements in them. If they did
not, they could be easily solved by local search methods. However it does not
necessarily follow that such problems can be solved by a GA that is good at
solving deceptive problems. The SBBH assumes that the misleading building
blocks will exist in the initial population, that they can be identified early in the
search before they are lost, and that the problem can be solved incrementally
by combining these building blocks, but perhaps the building blocks that have
misleading alternatives have little meaning until late in the search and so cannot
be expected to survive in the population.
Even if the SBBH turns out not to be as useful an hypothesis as originally
supposed, the increased propagation capabilities of pairwise mating may give a
GA (using pairwise mating) an advantage over a poolwise CCV algorithm. To
see why this is the case it is useful to define the prototypical individual for a
given population: for each locus we assign a one or a zero depending upon which
value is most frequent in the population (randomly assigning a value if they are
equally frequent). Suppose the population contains some maverick individual
that is quite far from the prototypical individual although it is near the optimum
(as measured by Hamming distance) but is of only average fitness. Since an
algorithm using a poolwise method of producing offspring will tend to produce
individuals that are near the prototypical individual, such an algorithm is unlikely
to explore the region around the maverick individual. On the other hand, a GA
using pairwise mating is more likely to explore the region around the maverick
individual, and so more likely to discover the optimum. Ironically, pairwise
mating is, in this respect, more mutation-like than poolwise mating. While
pairwise mating retains the benefits of CCV, it less subject to the majoritarian
tendencies of poolwise mating.
8.4 Representation
Although GAS typically use a bitstring representation, GAS are not restricted
to bitstrings. A number of early proponents of GAS developed GAS that use
other representations, such as real-valued parameters (Davis 1991 , Janikow
and Michalewicz I99 1, Wright 1991; see Chapter I6), permutations (Davis
1985, Goldberg and Lingle 1985, Grefenstette et a1 1985; see Chapter 17), and
treelike hierarchies (Antonisse and Keller 1987; see Chapter 19). Kozas genetic
programming (GP) paradigm (Koza 1992; see Chapter I I ) is a GA-based method
for evolving programs, where the data structures are LISP S-expressions, and
crossover creates new LISP S-expressions (offspring) by exchanging subtrees
TEAM LRN
from the two parents.
76
Genetic algorithms
77
suggested methods for allowing the GA to adapt its own coding. We noted
earlier that Holland proposed the inversion operator for rearranging the loci
in the string. Another approach to adapting the representation is Shaefer's
ARGOT system (Shaefer 1987). ARGOT contains an explicit parameterized
representation of the mappings from bitstrings to real numbers and heuristics
for triggering increases and decreases in resolution and for shifts in the ranges
of these mappings. A similar idea is employed by Schraudolph and Belew
( 1 992) who provide a heuristic for increasing the resolution triggered when the
population begins to converge. Mathias and Whitley ( 1994) have proposed
what they call delta coding. When the population converges, the numeric
representation is remapped so that the parameter ranges are centered around
the best value found so far, and the algorithm is restarted. There are also
heuristics for narrowing or extending the range.
There are also GAS with mechanisms for dynamically adapting the rate
at which CA operators are used or which operator is used. Davis, who has
developed a number of nontraditional operators, proposed a mechanism for
adapting the rate at which these operators are applied based on the past success
of these operators during a run of the algorithm (Davis 1987).
78
Genetic algorithms
8.6 Conclusion
Although the above discussion has been in the context of GAS as potential
function optimizers, it should be pointed out that Holland's initial GA work was
in the broader context of exploring GAS as adaptive systems (De Jong 1993).
GAS were designed to be a simulation of evolution, not to solve problems. Of
course, evolution has come up with some wonderful designs, but one must not
lose sight of the fact that evolution is an opportunistic process operating in an
environment that is continuously changing. Simon has described evolution as
a process of searching where there is no goal (Simon 1983). This is not to
question the usefulness of GAS as function optimizers, but only to emphasize
that the perspective of function optimization is somewhat different from that of
adaptation, and that the requirements of the corresponding algorithms will be
somewhat different.
References
Antonisse H J and Keller K S 1987 Genetic operators for high-level knowledge
representations Pro(.. 2nd ltit. Coiij; on Genetic Algorithms (Cunzhriclge,MA, 1987)
cd J J Grefenstette (Hillsdale, NJ: Erbaum) pp 69-76
Back T. Hoffmeister F and Schwefel H 1991 A survey of evolution strategies Pro(..
4th lnt. Coi$ on Genetic A1goritliriz.s (San Diego, CA, 1991) ed R K Belew and
L B Booker (San Mateo, CA: Morgan Kaufmann) pp 2-9
Baluja S 1995 A n Empiricd Cornparison of Se~*enItt#rutiL'earid Evolictionan Function
Optinziztion Hei4ristic.s Carnegie Mellon University School of Computer Science
Technical Report CMU-CS-95- I93
Caruana R A and Schafler J D 1988 Representation and hidden bias: Gray vs. binary
coding for genetic algorithms Proc. 5th ltzt. Conj: oii Machiize Leuniing (San Mateo,
CA: Morgan Kaufmann) pp 153-61
Davis L 1985 Applying adaptive algorithms to epistatic domains Pruc. I n t . Joirit
Coifiretice o i i Artijcicil ltitelligetice pp 162-4
__ 1987 Adaptive operator probabilities in genetic algorithms Proc. 3rd lnt. Conj: o t i
Genetic. Algorithms (F(iirjiix, VA, I989) ed J D Schaffer (San Mateo, CA: Morgan
Kaufmann) pp 61-9
-199 I Hybridization and numerical representation The HuncIbook of Genetic
Algorithms ed L Davis (New York: Van Nostrand Reinhold) pp 61-71
De Jong K I975 A n Ancilyis of the Behuviar of N Clu.s.s cf Genetic Aduptiiv Swtetns
Doctoral Thesis, Department of Computer and Communication Sciences, University
of Michigan
-1993 Genetic algorithms are not function optimizers Foutzdutions of Genetic
,4Igorithtrz.r 2 ed D Whitley (San Mateo, CA: Morgan Kaufmann) pp 5-17
De Jong K and Sarma J 1993 Generation gaps revisited Founddons of Genetic
Algorithms 2 ed D Whitley (San Mateo, CA: Morgan Kaufmann) pp 19-28
Eshelman L J I991 The CHC adaptive search algorithm: how to have safe search
when engaging in nontraditional genetic recombination Foundutions oJ' Genetic
TEAM LRN
Algorithnzs ed G J E Rawlins (San Mateo, CA: Morgan Kaufmann) pp 265-83
References
79
Eshelman L J and Schaffer J D 1993 Real-coded genetic algorithm4 and interval schemata
Fourzdatioiis of Genetic Algorithnzs 2 ed D Whitlcy (San Matco. CA: Morgan
Kaufmann) pp 187-202
-I
995 Productive recombination and propagating and pre4ening xhemata
Foitiickltioiis of Genetic Algorithms 3 ed D Whitley (San Mateo, CA: Morgan
Kaufmann) pp 299-3 13
Goldberg D E 1987 Simple genetic algorithms and the minimal, deceptive problcm
Genetic Algorithms (xiid Sirnitlated Aniieciling cd L Datri4 (San Mateo, CA: Morgan
Kaufmann) pp 74-88
-I989 Genetic Algorithins in Seurch, Optiinixition, tiiirl M(ichine kririziiig (Reading,
MA: Addiwn-Wc4ey)
Goldberg D E and Deb K 1991 A comparative analysi\ of \election wherne9 used in
genetic algorithms Foitizdiitions c~ Genetic. Algoritliiii, ed G J E Rawlins (San
Mateo, CA: Morgan Kaufmann) pp 69-93
Goldberg D E, Deb K, Kargupta H and Harik G 1993 Rapid, accurate optimiration
of difficult problems using fast messy genetic algorithms Proc. -5th I i i t . Cont. oii
Genetic Algoritlinis (Urbaiza-Chuinpai~n,IL, 1993) cd S Forrest (San Matco. CA:
Morgan Kaufmann) pp 56-64
Goldberg D E, Deb K and Korb B 1991 Dont worry, be mc\sy Proc. 4th I i i t . Coiif. o i i
Genetic. Algorithms ( S m Diego, CA, 1991) ed R K Bclew and L B Booher (San
Mateo, CA: Morgan Kaufmann) pp 24-30
Goldberg D E and Lingle R L 1985 Alleles, loci, and the traveling salesman problcm
Proc. 1st Int. Car$ 011 Cerietic Algorithms (Pittshur~qh,PA, 1085) cd J J Grefcmtette
(Hillsdale, NJ: Erbaum) pp 154-9
Cordon V S and Whitley 1993 Serial and parallel genetic algorithm\ and function
optimizers Proc. 5th Int. Conf on Genetic Algorithm 5 ( Ur~~rriici-CIi~ni~~trigrl.
IL.
1993) ed S Forrest (San Mateo, CA: Morgan Kaufmann) pp 177-83
Grefenstette J J 1 993 Deception considered harmful FoitiickItioiis of Geiwtic AlCqoritlims
2 ed D Whitley (San Mateo, CA: Morgan Kaufmann) pp 75-91
Grefenstette J J, Gopal R, Roqmaita B J and Van Gucht D 1985 Genetic algorithm4 for thc
traveling salesman problem Proc. 1st Itit. Co@ oii Griwtic Al~qorithins( PittshicrCqh,
PA, 1985) ed J J Grefenstette (Hillsdale, NJ: Erbaum) pp 160-8
Holland J H 1975 Aduptatiori iri Natiirctl and Artijcicil Sl-steim (Ann Arbor, MI:
University of Michigan Press)
Janikow C 2 and Michalewicz Z 1991 An experimental comparison of binary and
floating point representations in genetic algorithms Proc. 4th Int. Coi!f: on Geiietic
Algorithms (Sciiz Diego, CA, 1991) ed R K Belew and L B Booker (San Mateo. CA:
Morgan Kaufmann) pp 3 1-6
Koza J I 992 Geizetic Prograinnzing: on the Progrciininiiig of Cotnpicters I n 7 M t w i i s of
Naturd Selection mid Genetics (Cambridge, MA: MIT Pres5)
Liepins G E and Vose M D 1991 Representational issues in genetic optimimtion J. E i p .
Tlieor. AI 2 101-15
Mathias K E and Whitley L D 1994 Changing representations during \earth: a
comparative study of delta coding E\~olutioiiur;~~
Conipirt. 2
Miihlenbein H and Schlierkamp-Voosen I993 The scicncc of brceding and its application
TEAM LRN Conipict. 1
to the breeder genetic algorithm Evolictiorzan
Genetic algorithms
80
Radcliffe N J I991 Forma analysis and random respectful recombination Proc.. 4th / t i t .
Corzj: ot1 Gerietic Algoritht?i.s(Sun Diego, CA, 1991) ed R K Belew and L B Booker
(San Mateo, CA: Morgan Kaufmann) pp 222-9
Schaffer J D, Eshelman L J and Offutt D 1991 Spurious correlations and premature
convergence in genetic algorithms Foutidations c.$ Genetic Algorithms ed
G J E Rawlins (San Mateo, CA: Morgan Kaufmann) pp 102-12
Schraudolph N N and Belew R K 1992 Dynamic parameter encoding for genetic
algorithms Mwhitie Learning 9 9-2 1
Shaefer C G I987 The ARGOT strategy: adaptive representation genetic optimizer
technique Genetic- Algorithnis cind Their Applictitiot2.s: Proc. 2nd Itit. Cotif: oti
Genetic Algorithms { Cmihridge, MA, I987) ed J J Grefenstette (Hillsdale, NJ:
Erlbaum) pp 50-8
Simon H A 1983 Rectson in Hitmm Aflciirs (Stanford, CA: Stanford University Press)
Spears W M and De Jong K A 1991 On the virtues of parameterized uniform crossover
Proc. 4th Itit. Cot$ on Genetic Algorithms (San Diego, CA, 1991) ed R K Belew
and L B Booker (San Mateo, CA: Morgan Kaufmann) pp 230-6
Syswerda G 1989 Uniform crossover in genetic algorithms Proc. 3rd Int. Cot$ 011
Generic A1gorirhni.s (Elctitfkr. VA, 1989) ed J D Schaffer (San Mateo, CA: Morgan
Kaufmann) pp 2-9
-199I Schedule optimization using genetic algorithms Hutidhook oj Genetic
Algoritlrt~ised L Davis (New York: Van Nostrand Reinhold) pp 3 3 2 4 9
-I993 Simulated crossover in genetic algorithms Foittid~ition,sof Genetic Algorithms
2 ed D Whitley (San Mateo. CA: Morgan Kaufniann) pp 239-55
Whitley D I989 The GENITOR algorithm and selection pressure: why rank-based
allocation of reproductive trials is best Proc. 3rd Itit. Cot$ on Genetic Algorithm.s
(F(iitj+&r.VA, 1989) ed J D Schaffer (San Mateo, CA: Morgan Kaufmann) pp 1 16-2 1
Whitley D, Starkweather T and Fuquay D 1989 Scheduling problems and traveling
salesmen: the genetic edge recombination operator Pro(.. 3rd Itit. Conf: oti
Genetic Algorithnis (FuitjL, VA, 1989) ed J D Schaffer (San Mateo, CA: Morgan
Kaufmann) pp 116-21
Wright A 1991 Genetic algorithms for real parameter optimitation Foiitirf&ms ~ f G e t i e t i ~
Al,gorithms ed G J E Rawlins (San Mateo, CA: Morgan Kaufmann) pp 205-18
TEAM LRN
Evolution strategies
Giinter Rudolph
81
82
Evolution strategies
is probably due to the fact that Rechenberg (1973) succeeded in analyzing the
simple version in Euclidean space with continuous mutation for several test
problems.
Within this setting the archetype of ESs takes the following form. An
individual a consisting of an element X E R is mutated by adding a normally
distributed random vector 2 N ( 0 , I t , ) that is multiplied by a scalar (T > 0 ( I t l
denotes the unit matrix with rank n ) . The new point is accepted if it is better than
or equal to the old one, otherwise the old point passes to the next iteration. The
selection decision is based on a simple comparison of the objective function
values of the old and the new point. Assuming that the objective function
f : R + R is to be minimized, the simple ES, starting at some point Xo E R,
is determined by the following iterative scheme:
(i)
First ideas to extend the simple ES (9.1) can be found in the book by Rechenberg
(1973, pp 78-86). The population consists of / A > I parents. Two parents are
selected at random and recombined by multipoint crossover and the resulting
individual is finally mutated. The offspring is added to the population. The
selection operation chooses the I-( best individuals out of the I-(
1 in total to
serve as parents of the next iteration. Since the search space was binary, this ES
was exactly the same evolutionary algorithm as became known later under the
term steady-state genetic algorithm (Section 28. I ) . The usage of this algorithmic
scheme for Euclidean search spaces poses the problem of how to control the
step length control parameter 0,.Therefore, the steady-state ES is no longer
TEAM LRN
in use.
83
+
+
(9.2)
r=l
]=/+I
Evolution strategies
84
(i)
Let w E (0, 2nj- denote the angles that are necessary to build the
orthogonal rotation matrix T(w) via (9.2). The mutated angles o:+)~are
obtained by
( 1 )
(1)
= (w,
q Z : i ) mod 2n
According to Schwefel (1995) a good heuristic for the choice of the constants
appearing in the above mutation operation is
but recent extensive simulation studies (Kursawe 1996) revealed that the above
recommendation is not the best choice-especially in the case of multimodal
objective functions i t seems to be better to use weak selection pressure ( p / A not
too small) and a parametrization obeying the relation r > 11. As a consequence,
a final recommendation cannot be TEAM
given LRN
here, yet.
85
+ O (O+ w ) mod 4n
ux + (I - U)X,
2
2
0
where I is the unit matrix and U is a random diagonal matrix whose diagonal
entries are either zero or one with the same probability. Note that the angles
must be adjusted to the interval ( 0 , 2 n ] .
After these preparations a sketch of a contemporary ES can be presented:
Generate ,U initial parents of the type (X,0 , w ) and determine their objective
function values f ( X ) .
repeat
do A times:
Choose p 2 2 parents at random.
Recombine their angles, standard deviations, and object variables.
Mutate the angles, standard deviations, and object variables of the
preliminary offspring obtained via recombination.
Determine the offsprings objective function value.
Put the offspring into the offspring population.
end do
Select the ,U best individuals either from the offspring population
or from the union of the parent and offspring population.
The selected individuals represent the new parents.
until some stopping criterion is satisfied.
It should be noted that there are other proposals to adapt u l . In the case
of a ( I , A ) ES with A = 3 k and k E N, Rechenberg (1994, p 47) devised
the following rule: Generate k offspring with U ~ k, offspring with C O , and k
offspring with U , / C for some c > 0 ( c = 1.3 is recommended for I I 5 100, for
larger n the constant c should decrease).
Further proposals, that are however still in an experimental state, try
to derandomize the adaptation process by exploiting information gathered in
preceding iterations (Ostermeier et al 1995). This approach is related to
(deterministic) variable metric (or quasi-Newton) methods, where the Hessian
matrix is approximated iteratively by certain update rules. The inverse of the
Hessian matrix is in fact the optimal choice for the covariance matrix C. A
TEAM
large variety of update rules is given
by LRN
the Oren-Luonherger c-lnss (Oren and
86
Evolution strategies
9.3. I
Herdy (1992) used A subpopulations, each of them possessing its own different
and fixed step size 0 . Thus, there is no step size control at the level of
individuals. After y generations the improvements (in terms of fitness) achieved
by each subpopulation is compared to each other and the best p subpopulations
are selected. Then the process repeats with slightly modified values of 0.Since
subpopulations with a near-optimal step size will achieve larger improvements,
they will be selected (i.e. better step sizes will survive), resulting in an alternative
method to control the step size. TEAM LRN
87
References
.EY
g(x) = max{f(x. y ) : y E Y ) .
References
Herdy M 1992 Reproductive isolation as strategy parameter in hierachically organized
evolution strategies Parallel Problem Solving from Nutirrt?, 2 ( P m c . 2nd lrrt. Conf
on Parallel Prohlrrn Sohling from Nature, Brus.sel,s, 1992) ed R Manner and B
LRN7
Manderick (Amsterdam: Elsevier)TEAM
pp 207-1
88
Evolution strategies
Klockgether J and Schwefel H-P I970 Two-phase nozzle and hollow core jet experiments
Pro(.. I l t h Synip. on Etigineering Aspect\ of Mrigrietoh?.drod?.nclrrtic.Jed D Elliott
(Pasadena, CA: California lnstitute of Technology) pp 141-8
Kursawe F 1996 Breeding evolution strategies-first results, talk presented at Dagstuhl
lectures Applicutioiis o j E\dittioiiup Algorithnt Y (March 1996)
Lohmann R I992 Structure evolution and incomplete induction Purullel Problem Sol\-ing
frotti Nuture, 2 (Pro(*.2nd Int. Conj: on Parullel Problem Sohirig from Nuture,
Brusseis, l Y Y 2 ) ed R Manner and B Manderick (Amsterdam: Elsevier) pp 175-85
Oren S and Luenberger D 1974 Self scaling variable metric (SSVM) algorithms. Part 11:
criteria and wfficient conditions for scaling a class of algorithms Muntigemettt SCI.
20 845-62
O\termeier A. GawelcLyk A and Hansen N 1995 A derandomked approach to \elfadaptation of evolution strategies Evolut. Conzpict. 2 369-80
Rechenberg I I965 Cybernetic solution path of un expeririientcil problem Library
Translation 1 122, Royal Aircraft Establishment, Farnborough, UK
-1913 E~~olittions.ctrcitegie:Optiniierurig techni,wher Systeme nuch Principien der
hiologi.\ (*henE\dution (Stuttgart: Frommann- Hol~boog
)
-1978 Evolutionsstrategien Siti~ulatiori.~riictl~oderi
in dcr Medicin und Biologic ed
B Schneider and U Ranft (Berlin: Springer) pp 83-1 14
-1994 E~v)lirtioii.\.\trtite~ie
Y4 (Stuttgart: Frommann-Holzboog)
Rudolph G 1992 On correlated mutations in evolution strategies Purullel Prohlerzi Soliirig
jrom N~iture,2 (Pro(.. 2nd lnt. Car$ o)i Pm-ullel Problem Solving from Nutirrr,
Bnr.\wls, 1992) ed R Manner and B Manderich (Amsterdam: Elsevier) pp 105-14
Schwefel H-P I965 Kyherrietische Elwlutioii cils Strutegie der e.~~~eritiir,itelle~i
Forsc-hung
in der Striit~iiing.\trc.hnik Diplomarbeit, Technical University of Berlin
-1977 Nunierische Optiniierurig \or1 Compirter-Modrllrn niittel\ der E\~olirtion.s,,triitegie (Base!: Birkhauser)
-1995 E\dirtion untl Optitnunz Seeking (New York: Wiley )
Schbefel H-P and Rudolph G 1995 Contemporary evolution strategies Ad\wic*e\ in
Artifkicil Ljfe ed F Morana et cil (Berlin: Springer) pp 893-907
Sebald A V and SchlenLig J 1994 Minimax design of neural net controllers for highly
uncertain plants IEEE Trtins. Neirrul Netirw-ks - 5 73-82
TEAM LRN
10
Evolutionary programming
V William Porto
10.1 Introduction
Evolutionary programming (EP) is one of a class of paradigms for simulating
evolution which utilizes the concepts of Darwinian evolution to iteratively
generate increasingly appropriate solutions (organisms) in light of a static or
dynamically changing environment. This is in sharp contrast to earlier research
into artificial intelligence research which largely centered on the search for
simple heuristics. Instead of developing a (potentially) complex set of rules
which were derived from human experts, EP evolves a set of solutions which
exhibit optimal behavior with regard to an environment and desired payoff
function. In a most general framework, EP may be considered an optimization
technique wherein the algorithm iteratively optimizes behaviors, parameters, or
other constructs. As in all optimization algorithms, it is important to note that
the point of optimality is completely independent of the search algorithm, and
is solely determined by the adaptive topography (i.e. response surface) (Atmar
1992).
In its standard form, the basic evolutionary program utilizes the four main
components of all evolutionary computation (EC) algorithms: initialization,
variation, evaluation (scoring), and selection. At the basis of this, as well as
other EC algorithms, is the presumption that, at least in a statistical sense,
learning is encoded phylogenically versus ontologically in each member of
the population. Learning is a byproduct of the evolutionary process as
successful individuals are retained through stochastic trial and error. Variation
(e.g. mutation) provides the means for moving solutions around on the search
space, preventing entrapment in local minima. The evaluation function directly
measures fitness, or equivalently the behavioral error, of each member in the
population with regard to the environment. Finally, the selection process
probabilistically culls suboptimal solutions from the population, providing an
efficient method for searching the topography.
The basic EP algorithm starts with a population of trial solutions which are
initialized by random, heuristic, or other appropriate means. The size of the
TEAM LRN
89
90
Evolutionary programming
population, p , may range over a broadly distributed set, but is in general larger
than one. Each of these trial solutions is evaluated with regard to the specified
fitness function. After the creation of a population of initial solutions, each
of the parent members is altered through application of a mutation process;
in strict EP, recombination is not utilized. Each parent member i generates
A, progeny which are replicated with a stochastic error mechanism (mutation).
The fitness or behavioral error is assessed for all offspring solutions with the
selection process performed by one of several general techniques including: (i)
the best p solutions are retained to become the parents for the next generation
(elitist, see Section 28.4), or (ii) p of the best solutions are statistically retained
( tournament, see Chapter 24), or (iii) proportional-based selection (Chapter 23).
In most applications, the size of the population remains constant, but there is no
restriction in the general case. The process is halted when the solution reaches
a predetermined quality, a specified number of iterations has been achieved, or
some other criterion (e.g. sufficient convergence) stops the algorithm.
EP differs philosophically from other evolutionary computational techniques
such as genetic algorithms (GAS) (Chapter 8) in a crucial manner. EP is a
top-down versus bottom-up approach to optimization. It is important to note
that (according to neo-Darwinism) selection operates only on the phenotypic
expressions of a genotype; the underlying coding of the phenotype is only
affected indirectly. The realization that a sum of optimal parts rarely leads
to an optimal overall solution is key to this philosophical difference. GAS
rely on the identification, combination, and survival of good building blocks
(schemata) iteratively combining to form larger better building blocks. In a
GA, the coding structure (genotype) is of primary importance as it contains
the set of optimal building blocks discovered through successive iterations.
The building block hypothesis is an implicit assumption that the fitness is a
separable function of the parts of the genome. This successively iterated local
optimization process is different from EP, which is an entirely global approach
to optimization. Solutions (or organisms) in an EP algorithm are judged solely
on their fitness with respect to the given environment. No attempt is made
to partition credit to individual components of the solutions. In EP (and in
evolution strategies (ESs), see Chapter 9), the variation operator allows for
simultaneous modification of all variables at the same time. Fitness, described
in terms of the behavior of each population member, is evaluated directly, and is
the sole basis for survival of an individual in the population. Thus, a crossover
operation designed to recombine building blocks is not utilized in the general
forms of EP.
10.2
History
History
91
1/6
Figure 10.1. A simple finite-state machine diagram. Input symbols are shown to the
left of the slash. Output symbols are to the right of the slash. The finite-state machine
is presumed to start in state A.
92
Evolutionary programming
Fogel (see Fogel 1964, Fogel et crl 1966) used EP on a series of successively
more difficult prediction tasks. These experiments ranged from simple twoTEAM cyclic
LRN sequences degraded by addition
symbol cyclic sequences, eight-symbol
History
93
20
40
60
80
100 120
140
160
Number of symbols experienced
180
200
220
94
Evolutionary programming
numbers, five were missed, but of the next 65 primes, none were missed. Fogel
et crl (1966) indicated that the machines demonstrated the capability to quickly
recognize numbers which are divisible by two and three as being nonprime,
and that some capability to recognize divisibility by five as being indicative
of nonprimes was also evidenced. Thus, the machines generated evidence of
learning a definition of primeness without prior knowledge of the explicit nature
of a prime number, or any ability to explicitly divide.
Fogel and Burgin ( 1969) researched the use of EP in game theory. In
a number of experiments, EP was consistently able to discover the globally
optimal strategy in simple two-player, zero-sum games involving a small number
of possible plays. This research also showed the ability of the technique to
outperform human subjects in more complicated games. Several extensions were
made to the simulations to address nonzero-sum games (e.g. pursuit evasion.)
A three-dimensional model was constructed where EP was used to guide an
interceptor towards a moving target. Since the target was, in most circumstances,
allowed a greater degree of maneuverability, the success or failure of the
interceptor was highly dependent upon the learned ability to predict the position
of the target without CI priori knowledge of the targets dynamics.
A different aspect of EP was researched by Walsh et er1 (1970) where EP
was used for prediction as a precursor to automatic control. This research
concentrated on decomposing a finite-state machine into submachines which
could be executed in parallel to obtain the overall output of the evolved system.
A primary goal of this research was to maximize parsimony in the evolving
machines. In these experiments, finite-state machines containing seven and
eight states were used as the generating function for three output symbols. The
performance of three human subjects was compared to the evolved models when
predicting the next symbol in the respective environments. In these experiments,
EP was consistently able to outperform the human subjects.
TEAM LRN
History
95
where:
U ' is an individual member in the population
p 3 I is the size of the parent population
h > 1 is the size of the offspring population
P ( t ) := { a ; ( t ) ,ah(t). . . . , ~ ~ ~ is
( tthe) population
)
at time t
CD : Z -+ '31 is the fitness mapping
m(..),,,
is the mutation operator with controlling parameters
.s(-), is the selection operator 3 s(-), : (Z* U I / ' + * ) + 1''
Q E {Vr, P ( t ) } is a set of individuals additionally accounted for in the
selection step, i.e. parent solutions.
Other than on initialization, the search space is generally unconstrained;
constraints are utilized for generation and initialization of starting parent
solutions. Constrained optimization may be addressed through imposition of
the requirement that (i) the mutation operator (Section 32.4) is formulated to
only generate legitimate solutions (often impossible) or ( i i ) a penalty function
is applied to offspring mutations lying outside the constraint bounds in such
a manner that they do not become part of the next generation. The objective
function explicitly defines the fitness values which may be scaled to positive
values (although this is not a requirement, it is sometimes performed to alter
the range for ease of implementation).
In early versions of EP applied to continuous parameter optimization (Fogel
1992) the mutation operator is Gaussian with a zero mean and variance obtained
for each component of the object variable vector as the square root of a linear
transform of the fitness value cp(z).
x, ( k
+ 1 ) :=
Xf ( k )
96
Evolutionary programming
.r,(k
+ 1) := . r , ( k ) + u , ( k )t N,(O,1)
+
~ , ( k 1 ) := ~ , ( k ) [ a u , ( k ) ] * N,(O,1).
The variable (T ensures that the variance U , remains nonnegative. Fogel
(1992) suggests a simple rule wherein V u , ( k ) 5 0, u , ( k ) is set to 6, a value
close to but not identically equal to zero (to allow some degree of mutation).
The sequence of updating the object variable .U, and variance U , was proposed to
occur in opposite order from that of ESs (Back and Schwefel 1993, Rechenberg
1965, Schwefel 1981). Gehlhaar and Fogel (1096) provide evidence favoring
the ordering commonly found in ES.
Further development of this theme led Fogel (1991a, 1992) to extend the
procedure to alter the correlation coefficients between components of the object
vector. A symmetric correlation coefficient matrix P is incorporated into
the evolutionary paradigm in addition to the self-adaptation of the standard
deviations. The components of P are initialized over the interval [ - 1 , 11
and mutated by perturbing each component, again, through the addition of
independent realizations from a Gaussian random distribution. Bounding
limits are placed upon the resultant mutated variables wherein any mutated
coefficient which exceeds the bounds [ - I . I ] is reset to the upper or loher
limit, respectively. Again, this methodology is similar to that of Schwefel
(1981 ), as perturbations of both the standard deviations and rotation angles
(determined by the covariance matrix P) allow adaptation to arbitrary contours
TEAM LRN
on the error surface. This self-adaptation
through the incorporation of correlated
Current directions
97
98
Evolutionary programming
cull lower-scoring members from the population. Optimization of the tours was
quite rapid. In one such experiment with 1000 cities uniformly distributed, the
best tour (after only 4 x 10 function evaluations) was estimated to be within
5-75? of the optimal tour length. Thus, excellent solutions were obtained after
searching only an extremely small portion of the total potential search space.
EP has also been utilized in a number of medical applications. For
example, the issue of optimizing drug design was researched by Gehlhaar P t
ul (1995). EP was utilized to perform a conformational and position search
within the binding site of a protein. The search space of small molecules
which could potentially dock with the crystallographically determined binding
site was explored iteratively guided by a database of crystallographic proteinligand complexes. Geometries were constrained by known physical ( i n three
dimensions) and chemical bounds. Results demonstrated the efficacy of this
technique as it was orders of magnitude fater in finding suitable ligands than
previous hands-on methodologies. The probability of successfully predicting the
proper binding modes for these ligands was estimated at over 95% using nominal
values for the crystallographic binding mode and number of docks attempted.
These studies have permitted the rapid development of several candidate drugs
which are currently in clinical trials.
The issue of utilizing EP to control systems has been addressed widely
(Fogel and Fogel 1990, Fogel 1991a, Page P t cil 1992, and many others).
Automatic control of fuzzy heating, ventilation, and air conditioning (HVAC)
controllers was addressed by Haffner and Sebald (1993). In this study, a
nonlinear, multiple-input. multiple-output (MIMO) model of a HVAC system
was used and controlled by it fuzzy controller designed using EP. Typical fuzzy
controllers often use trial and error methods to determine parameters and transfer
functions, hence they can be quite time consuming with a complex MIhlO
HVAC system. These experiments used EP to design the membership functions
and values (later studies were extended to find rules as well a s responsibilities
of the primary controller) to automate the tuning procedure. EP worked in
an overall search space containing 76 parameters, 10 controller inputs, seven
controller outputs. and 80 rules. Simulation results demonstrated that EP was
quite effective at choosing the membership functions of the control laboratory
and corridor pressures in the model. The synergy of combining EP with fuzzy
set constructs proved quite fruitful in reducing the time required to design a
stable, functioning HVAC system.
Game theory has always been at the forefront of artificial intelligence
research. One interesting game, the iterated prisoners dilemma, has been
studied by numerous investigators (Axelrod 1987. Fogel I99 I b, Harrald and
Fogel 1996, and others). In this two-person game, each player can choose
one of two possible behavioral policies: defection or cooperation. Defection
implies increasing ones own reward at the expense of the opposing player,
while cooperation implies increasing the reward for both players. If the game
LRN move is defection. If the players
is played over a single iteration, theTEAM
dominant
99
Current direction s
In addition, the payoff matrix defining the game is subject to the following
constraints (Rapoport 1966):
Both neural network approaches (Harrald and Fogel 1996) and finite-state
machine approaches (Fogel 1991b) have been applied to this problem. Finitestate machines are typically used where there are discrete choices between
cooperation and defection. Neural networks allow for a continuous range of
choices between these two opposite strategies. Results of these preliminary
experiments using EP, in general, indicated that mutual cooperation is more
likely to occur when the behaviors are limited to the extremes (the finitestate machine representation of the problem), whereas in the neural network
continuum behavioral representation of the problem, i t is easier to slip into a
state of mutual defection.
Development of interactively in tell igen t be haviors was investigated by Foge 1
et a1 (1996). EP was used to optimize computer-generated force (CGF)
behaviors such that they learned new courses of action adaptively as changes
in the environment (i.e. presence or absence of opposing side forces) were
encountered. The actions of the CGFs were created at the response of an event
scheduler which recognized significant changes in the environment as perceived
by the forces under evolution. New plans of action were found during these
event periods by invoking an evolutionary program. The iterative EP process
TEAM
LRNmet, and relinquished control of the
was stopped when time or CPU limits
were
100
Evolutionary programming
simulated forces back to the CGF simulator after transmitting newly evolved
instruction sets for each simulated unit. This process proved quite successful
and offered a significant improvement over other rule-based systems.
References
Aho A V, Hopcroft J E and Ullman J D 1974 Tlic Desigii trricl Aticrlv.\i,\ of Cotnpiitrr
Algorithms (Reading, MA: Addison-Wesley) pp 143-5, 3 18-26
Angeline P, Saunders G and Pollack J 1994 Complete induction of recurrent neural
networks Proc. 3rd Anti. Conj: on E\vliitiotitiry Progrcirtirnitig (Sirti Diego, C A , 1994)
ed A V Sebald and L J Fogel (Singapore: World Scientific) pp 1-8
Atmar W 1992 On the rules and nature of simulated evolutionary programming Pro(*.
1st Anti. Cot$ o t i E\*oliitioticic Progrtittiniitig (LaJollu, C A , 1992) ed D B Fogel
and W Atmar (San Diego, CA: Evolutionary Programming Society) pp 17-26
Axelrod R 1987 The evolution of strategies in the iterated prisoners dilemma Griirrii.
Algorirhms mid Siriiiiltitetl Atirieciling ed L DaL is (London) pp 32-42
Back T and Schwefel H-P 1993 An overview of evolutionary algorithms for parameter
optimimtion E\*oliitioncin C(itnpit. 1 1-23
Brotherton T W and Simpson P K 1095 Dynamic feature set training of neural
networks for classification E\diitionciry Progrutt?itnitig IV: Proc. 4th Anti. Cot!/.o t i
E\vliitioticin Progrtrtnmirig (Strti Diego, C A , l W 5 ) ed J R McDonnell, R G Reynold\
and D B Fogel (Cambridge, MA: MIT Press) pp 83-94
Burton D M 1976 E l e m e n t q v Nirriiher the or^ (Boston, MA: Allyn and Bacon) p 136-52
Flood M M 1962 Stochastic learning theory applied to choice experiment\ with cats,
dogs and men Buhmiorcrl Sci. 7 289-3 14
Fogel D B 1988 An evolutionary approach to the traveling salesman problem Biol.
Cyhumet. 60 139-44
-I99 1 a System Identificutioti tlzroiigh Simulated E\*oliitioti (Needham. MA: Ginn)
---I
99 1 b The evolution of intelligent decision making in gaming C\hrrnut. LY~:\t 22
223-36
TEAM
-1992 E\*ol\+ig Artijrticil Ititrlligrtii~e
PhDLRN
Dissertation, Univer\ity o f California
References
101
102
Evolutionary programming
Further reading
There are several excellent general references available to the reader interested
in furthering his or her knowledge in this exciting area of EC. The following
books are a few well-written examples providing a good theoretical background
in EP as well as other evolutionary algorithms.
1 . Back T 1996 E\wlictionciry Algorithms in Theory c i n d Prcictice (New York: Oxford
University Press)
2. Fogel D B I995 E\diitionury Compiitcition, Tmrurcl ci N e ~ vPhilosophy ($Machine
lntelligenc*e (Piscataway, NJ: IEEE)
(.f Computer
Models (Chichester:
4. Schwefel H-P I995 Eidution cind Optirnizition Seeking (New York: Wiley)
TEAM LRN
11
Derivative methods in genetic
programming
Kenneth E Kinnear, Jr
11.1 Introduction
This chapter describes the fundamental concepts of genetic programming (GP)
(Koza 1989, 1992). Genetic programming is a form of evolutionary algorithm
which is distinguished by a particular set of choices as to representation,
genetic operator design, and fitness evaluation. When examined in isolation,
these choices define an approach to evolutionary computation (EC) which is
considered by some to be a specialization of the genetic algorithm (GA). When
considered together, however, these choices define a conceptually different
approach to evolutionary computation which leads researchers to explore new
and fruitful lines of research and practical applications.
103
I 04
105
are functions of two inputs, and one is a function of one input. Each produces
a single output and no side effect.
The fitness evaluation for this particular indi\liditril is determined by the
effectiveness with which it will produce the correct logical output for all of the
test cases against which i t is tested.
One way to characterize the design of a representation for an application
of genetic programming to a particular problem is to view it as the design of
a language, and this can be a useful point of view. Perhaps i t is more useful,
however, to view the design of a genetic programming representation as that
of the design of a virtual machine-since usually the execution engine must
be designed and constructed as well as the representation or language that is
executed.
The representation for the program (i.e. the definition of the functions and
terminals) must be designed along with the virtual machine that is to execute
them. Rarely are the programs evolved in genetic programming given direct
control of the central processor of a computer (although see the article by
Nordin (1994)). Usually, these programs are interpreted under control of a
virtual machine which defines the functions and terminals. This includes the
functions which process the data, the terminals that provide the inputs to the
programs, and any control functions whose purpose is to affect the execution
flow of the program.
As part of this virtual machine design task, it is important to note that the
output of any function or the value of any terminal may be used as the input to
any function. Initially, this often seems to be a trivial problem, but when actually
performing the design of the representation and virtual machine to execute that
representation, it frequently looms rather large. Two solutions are typically used
for this problem. One approach is to design the virtual machine, represented by
the choice of functions and terminals, to use only a single data type. In this way,
the output of any function or the value of any terminal is acceptable as input to
TEAM
LRN more than one data type to exist
any function. A second approach is
to allow
I06
in the virtual machine. Each function must then be defined to operate on any of
the existing data types. Implicit coercions are performed by each function on
its input to convert the data type that it receives to one that it is more normally
defined to process. Even after handling the data type problem, functions must
be defined over the entire possible range of argument values. Simple arithmetic
division must be defined to return some value even when division by zero is
attempted.
I t is important to note that the definition of functions and the virtual machine
that executes them is not restricted to functions whose only action is to provide
a single output value based on their inputs. Genetic programming functions
are often defined whose primary purpose is the actions they take by virtue of
their side-effects. These functions rnust return some value as well, but their real
purpose is interaction with an environment external to the genetic programming
system.
An additional type of side-effect producing function is one that implements
a control structure within the virtual machine defined to execute the genetically
evolved program. All of the common programming control constructs such as
if-then+lse, while-do, for, and others have been implemented as evolvable
control constructs within genetic programming systems. Looping constructs
must be protected in such a way that they will never loop forever. and usually
have an arbitrary limit set on the number of loops which they will execute.
As part of the initialization of a genetic programming run, a large number of
individual programs are generated at random. This is relatively straightforward,
since the genetic programming system is supplied with information about the
number of arguments required by each function, as well as all of the available
terminals. Random program trees are generated using this information, typically
of a relatively small size. The program trees will tend to grow quickly to be
quite large in the absence of some explicit evolutionary pressure toward small
size or some simple hard-coded lirnits to growth.
107
I08
parent tree. In practice it yields a offspring tree whose fitness has enough
relationship to that of its parents to support the evolutionary search process.
Variations in this crossover approach are easy to imagine, and are currently the
subject of considerable active research in the genetic programming community
(D'haeseleer 1994, Teller 1996).
Mutation (Chapter 3 2 ) is a genetic operator mhich can be applied to a \ingle
parent program tree to create an offspring tree. The typical mutation operator
used selects a point inside a parent tree, and generates a new random subtree
to replace the selected subtree. This random subtree i\ uwally generated by the
\ame procedure used to generate the initial population of program trees.
Finally, then, the last detailed distinction between genetic programming and a
more usual implementation o f the genetic algorithm is that of the assignment of
a fitness value for a individual.
I n genetic programming, the representation of the individual is a program
which. when executed under control of a defined virtual machine, implements
some algorithm. It may do this by returning some value (as would be the case
for a system t o learn a specitic Boolean function) or it might do this through the
performance of some task through the use of functions which have side-effects
that act on a simulated (or even the real) world.
The results of the program's execution are evaluated in some way, and this
evaluation represents the fitness of the individual. This fitness is used to drive
the selection process for copying into the next generation or for the selection of
parents to undergo genetic operations yielding offspring. Any selection operator
from those presented in Chapters 22-33 can be used.
There is certainly a desire to evolve programs using genetic programming
that are 'general', that is to say that they will not only correctly process the
fitness cases on which they are evolved, but will process correctly any fitness
cases which could be presented to them. Clearly, in the cases where there are
intinitely many possible cases, such as evolving a general sorting algorithm
(Kinnear 1993), the ekdutionary process can only be driven by a \ a y limited
number of fitness cases. Many of the lessons from machine learning on the
tradeoffs between generality and performance on training cases have been
helpful to genetic programming researchers, particularly those from decision
tree approaches to machine learning (Iba et (11 1994).
109
evolved, and are only considered executable when they are undergoing fitness
evaluation.
As genetic programming itself evolved in LISP, the programs that were
executed began to look less and less like LISP programs. They continued to be
tree structured but soon few if any of the functions used in the evolved programs
were standard LISP functions. Around I992 many people implemented genetic
programming systems in C and C++, along with many other programming
languages. Today, other than a frequent habit of printing the representation
of tree-structured genetic programs in a LISP-like syntax, there is no particular
connection between genetic programming and LISP.
There are many public domain implementations of genetic programming in
a wide variety of programming languages. For further details, see the reading
list at the end of this section.
110
This last example is a step along the path toward modeling embryonic
development in genetic programming. The opportunity exists to evolve
programs whose results are themselves programs. These resulting programs
are then executed and their values or side-effects are evaluated-and become
the fitness for the original evolving, program creating programs. The analogy
to natural embryonic development is clear, where the genetic material, the
genotype, produces through development a body, the phenotype, which then
either does or does not produce offspring, the fitness (Kauffman 1993).
Genetic programming is valuable in part because we find it natural to
examine issues such as those mentioned above when we think about evolutionary
TEAM LRN
computation from the genetic programming
perspective.
References
111
References
Angeline P J 1996 Two self-adaptive crossover operator\ for genetic programming
Adiunces in Genetic Prograrnrnirig 2 ed P J Angcline and K E Kinnear Jr
(Cambridge, MA: MIT Press)
Angeline P J and Pollack J B 1993 Competitive environments evolve better solutions for
complex tasks Ptm. 5th lnt. Con& on Genetic Algorithrm ( Uri~ctritr-Cliciiiii~tri~~ii,
IL,
July 1993) ed S Forrest (San Mateo, CA: Morgan Kaufmann)
Cramer N L 1985 A representation of the adaptive generation of simple sequential
programs Proc. 1st Int. Conf on Genetic Algorithm (Pittshi;qh, PA. J u l y 198-5)
ed J J Grefenstette (Hillsdale, NJ: Erlbaum)
Dhaeseleer P 1994 Context preserving crossover in genctic programming 1st IEEE Coiif
oti Eidutionun Coinpiitdon (Orlando, FL, June 1994) (Piscataway, NJ: IEEE)
Gruau F 1993 Genetic synthesis of modular neural networks PI-oc. 5th I i i t . Conf oti
Genetic Algorithins (Urbriiici-Chan2paign,IL, Jul! 1993) ed S Forrest (San Mateo,
CA: Morgan Kaufmann)
Holland J H 1975 Adrptation in Natural and ArtiJir-itrlS\,steins (Ann Arbor, MI: The
University of Michigan Press)
Iba H, de Garis H and Sato T 1994 Genetic programming using a minimum description
length principle Adimces in Genetic Progmtnniiiig ed K E Kinnear Jr (Cambridge.
MA: MIT Press)
Kauffman S A 1993 The Origins of Order: SelfOrg.citii~crtiori( i i i d Selwtioti iii E\dirtioii
(New York: Oxford University Press)
Kinnear K E Jr 1993 Generality and difficulty in genetic programming: evolving a sort
Proc. 5th Inr. Conf. on Genetic Algorithms ( ~ r h c l l i c l - ~ h c ~ n i ~ ~IL,
r r i July,
g n , 199-3) ed
S Forrest (San Mateo, CA: Morgan Kaufmann)
-I994 Alternatives in automatic function definition: a comparison of performance
A d i w i c ~ sin Genetic Programming ed K E Kinnear Jr (Cambridge. MA: MIT Pre\\)
Koza J R 1989 Hierarchical genetic algorithms operating on population\ of computer
programs Proc. I I th Int. Joint Cot$ on Artijkicil liitc4ligcwcx~(San Mateo. CA:
Morgan Kaufmann)
Nordin P 1994 A compiling genetic programming system that directly manipulates the
machine code Ad\mce.\ in Genetic ProRraniniing ed K E Kinnex Jr (Cambridge,
MA: MIT Press)
Perkis T 1994 Stack-based genetic programming Proc. 1st IEEE Irit. Conf o ~ i
Eivlutiotiar~Contputatioti (Orlando, FL, June 1994) (Piscataway, NJ: IEEE)
Reynolds C R 1994 Competition, coevolution and the game of tag Artijkitrl Ljfi IV:
Proc. 4th Int. Workshop on the Synthesis anti Sitnuldoii of Liiiiig Systetiis cd R A
Brooks and P Maes (Cambridge, MA: MIT Press)
Sims K 1994 Evolving 3D morphology and behavior by competition Artificitrl L i p I\/:
Proc. 4th Int. Workshop on the Synthesis urid Siinulntioti of Liiirig Sjstcwis ed R A
Brooks and P Maes (Cambridge, MA: MIT Press)
Teller A 1996 Evolving programmers: the co-evolution of intelligent recombination
operators Advance.\ in Genetic Prograrnrning 2 ed P J Angeline and K E Kinnear
Jr (Cambridge, MA: MIT Press) TEAM LRN
112
~~~~~
Further reading
1. Koza J R 1992 Genetic Progrcirnrnirig (Cambridge, MA: MIT Press)
The tirst book on the subject. Contains full instructions on the possible details of
carrying out genetic programming, as well as a complete explanation of genetic
algorithms (on which genetic programming is based). Also contains I I chapters
showing applications of genetic programming to a wide variety of typical artificial
intelligence, machine learning, and sometimes practical problems. Gives many
examples of how to design a representation of a problem for genetic programming.
3
Kom J R 1994 Genetic. Progrcrrtrtning / I (Cambridge, MA: MIT Press)
-.
4. Forrest S (ed) 1993 Proc. 5th Int. Cot$ on Genetic Algorithtns ( Urbuttu-C~z~it?lp~ii~tz,
IL. July 1993) (San Mateo, CA: Morgan Kaufmann)
Contains several interesting papers on genetic programming.
5. 1994 1st IEEE Cot$ on Evo/iitionur?. Compurution (Orfundo, FL, June 1994)
(Piscataway, NJ: IEEE)
Contains many papers on genetic programming as well as a wide assortment of
other EC-based papers.
6. Eshelman L J (ed) 1995 Proc-.6th lnt. Conj: on Genetic-Algorithms (Pittsburgh, PA,
July IYY.5) (Cambridge, MA: MIT Press)
Further read i n g
113
TEAM LRN
12
Learning classifier systems
Robert E Smith
12.1 Introduction
The learning classifier system (LCS) (Goldberg 1989, Holland et ul 1986) is
often referred to as the primary machine learning technique that employs genetic
algorithms (GAS). It is also often described as ii production system framework
with a genetic algorithm as the primary rule discovery method. However, the
details of LCS operation vary widely from one implementation to another. In
fact, no standard version of the LCS exists. In many ways, the LCS is more
of a concept than an algorithm. To explain details of the LCS concept, this
article will begin by introducing the type of machine learning problem most
often associated with the LCS. This discussion will be followed by a overview
of the LCS, in its most common form. Final sections will introduce the more
complex issues involved in LCSs.
TEAM LRN
115
I I6
action alters the probability of moving the plant from the current state to any
other state. Note that deterministic environments are a specific case. Although
this discussion will limit itself to discrete problems, most of the points made
can be related directly to continuous problems.
A characteristic of many reinforcement learning problems is that one may
need to consider a sequence of control actions and their results to determine
how to improve the controller. One can examine the implications of this by
associating a rerturd or cost with each control action. The error in state in
figure 12.3 can be thought of as a cost. One can consider the long-term effects
of an action formally as the expected, iriJitiite-lzori,7ctidiscmmted cvst:
h'c,
117
0 1 1 0 / 0 1 0
represents one of 16 states and one of eight actions. This string can also be
seen as a rule that says IF in state 0 1 1 0,THEN take action 0 1 0. In an
LCS, such a rule is called a classifier. One can easily associate a Q-value, or
other performance measures, with any given classifier.
Now consider generalizing over actions by introducing a dont care
character (#) into the state portion of a classifier. In other words, the string
# 1 1 # / 0 1 0
is a rule that says IF in state 0 1 1 0 OR state 0 1 1 1 OR state 1 1 1 0 OR
state 1 1 1 1, THEN take action 0 1 0.The introduction of this generality
allows an LCS to represent clusters of states and associated actions. By using
the genetic algorithm to search for such strings, one can search for ways of
clustering states together, such that they can be assigned joint performance
statistics, such as Q-values.
Note, however, that Q-learning is not the most common method of credit
assignment in LCSs. The most common method is called the bucket brigade
TEAM LRN
1 I8
There are two methods of using the genetic algorithm in LCSs. One is for each
genetic algorithm population member to represent an entire set of rules for the
problem at hand. This type of LCS is typified by Smiths LS-I which was
developed at the University of Pittsburgh. Often, this type of LCS is called
the Pitt approach. Another approach is for each genetic algorithm population
member to represent a single rule. This type of LCS is typified by the CS1 of Holland and Reitman (1978), which was developed at the University of
Michigan, and is often called the Michigan approach.
In the Pitt approach, crossover and other operators are often employed
that change the number of rules in any given population member. The Pitt
approach has the advantage of evaluating a complete solution within each
genetic algorithm individual. Therefore, the genetic algorithm can converge to a
homogeneous population, as in an optimization problem, with the best individual
located by the genetic algorithm search acting as the solution. The disadvantage
is that each genetic algorithm population member must be completely evaluated
as a rule set. This entails a large computational expense, and may preclude
on-line learning in many situations.TEAM LRN
119
In the 'Michigan' approach, one need only evaluate a single rule set, that
comprised by the entire population. However, one cannot use the usual genetic
algorithm procedures that will converge to a homogeneous population, since one
rule is not likely to solve the entire problem. Therefore, one must c.oe\wlr?ea
set of cooperative rules that jointly solve the problem. This requires a genetic
algorithm procedure that yields a diverse population at steady state, in a fashion
that is similar to sharing (Deb and Goldberg 1989, Goldberg and Richardson
1987), or other multimodal genetic algorithm procedures. I n some cases simply
dividing reward between similar classifiers that fire can yield sharing-like effects
(Horn et a1 1994).
12.5
As was noted earlier, the bucket brigade algorithm is the most common form of
credit allocation for LCSs. In the bucket brigade, each classifier has a strength,
S , which plays a role analogous to a @value. The bucket brigade operates as
follows:
(i)
(ii)
(iii)
(iv)
I20
In LCSs with internal messages, the bucket brigade can be used in its
original, explicit form. In this form, the next rule that acts is linked to the
previous rule through an internal message. Otherwise, the mechanics are similar
to those noted above. Once classifiers are linked by internal messages, they can
form rrdr drtrirzs that express complex sequences of actions.
12.7 Parasites
The possibility of rule chains introduced by internal messages, and by payback
credit allocation schemes such as the bucket brigade or @learning, also
introduces the possibility of rule parasites. Simply stated, parasites are rules that
TEAM LRN
in ;1 rule chain or a sequence of LCS
obtain fitness through their participation
121
actions, but serve no useful purpose in the problem environment. In some cases,
parasite rules can prosper, while actually degrading overall system performance.
A simple example of parasite rules in LCSs is given by Smith (1994). In this
study, a simple problem is constructed where the only performance objective is
to exploit internal messages as internal memory. Although fairly effective rule
sets were evolved in this problem, parasites evolved that exploited the bucket
brigade, and the existing rule chains, but that were incorrect for overall system
performance. This study speculates that such parasites may be an inevitable
consequence in systems that use temporal credit assignment (such as the bucket
brigade) and evolve internal memory processing.
122
References
Barto A G 1990 Soirie Lvtrrriirig Tusks jrom Li Control Perspec.ri\>eCOINS Technical
Report 90- 122, University of Massachusetts
Barto A G, Bradtke S J and Singh S P 1991 Rrcil-tiirir Leciriiirtg ciml Coiitml i r s i i i g
A.s~ti(~hro~ioii.s
Dyimriic- Progrtrmmiii~COINS Technical Report 9 1-57, University
of Massachusetts
Booker L B 1982 Intelligent behavior as an adaptation to the task environment
Dis.sertrition.sAhstnrc-ts lilt. 43 469B; University Microfilms 82 I4966
-1985 Improving the pertornmancc of genetic algorithms in classifier systems Prot..
Itit. Col$ 0 1 1 Genetic Aigorithms ciiid T h i r Applic~irionspp 80-92
-I989 Triggered rule discovery in classifier systems Proc. 3rd Int. Corzj: 0 1 1 G e i i i 4 .
Algorithms (Fairfus, VA, Jitiie 1989) ed J D Schaffer (San Mateo. CA: Morgan
Kaufmann) pp 265-74
Deb K and Goldberg D E I989 An investigation o f niche and species formation in genetic
function optimization Proc. 3rd hit, Coi$ oii Genetic Algoritkms (Fciirjh-, VA, Jirrir
1989) ed J D Schaffer (San Mateo, CA: Morgan Kaufmann) pp 32-50
Goldberg D E 1989 Gerietic A1gorithm.s in Selrrdi, Optiirii;.~itioii,ctiitl Mtrc-hine Leciriiiiig
(Reading, MA: Addison-Wesley)
Goldberg D E and Richardson J 1987 Genetic algorithms with sharing !or multimodal
function optimiLation Proc. 2nd Irit. Coi$ 0 1 1 Cmutic Algorithrrrs (Cm~ihritl~qr.
M A,
1987) ed J J Grefenstette (Hillsdale. NJ: Erlbuum) pp 41-9
Holland J H, Holyoak K J , Nisbett R E and Thagard P R 1986 Itidiic~tioii:Proc~c~.s.~c~.s
of
Iizfkreiicx~,Letrrtiing, mid D i . s c - o \ ~(Cambridge,
n~
MA: MIT Press)
Holland J H and Reitman J S 1978 Cognitive systems based on adaptive algori~hms
Ptrtterti Directed li!fereiit~rSystriiis ed D A Waterman and F Hayes-Roth (New
York: Academic) pp 3 13-24,
Horn J , Goldberg D E and Deb K I993 Implicit niching in a learning classifier system:
Nature's way E\dittioncir~ Coiripitt. 2 37-66
Riolo R L 1 986 C'FS-c':ti Pwktige of' Domuin I~rclept~~i~lrtit
Sirhroirtinrsfi)r lmpirtiieiitiri~~
Clnssifirr S!xtriirs ii r A rhitra n User-dejiirud E'ii ~iroiirrrcw
t s U n i ve rs i t y of M i c h i ga n ,
Logic of Computers Croup, Technical Report
Robertson G G and Riolo R 1988 A tale of two classifier systems Mnchine Lwriiiii,q 3
TEAM LRN
139-60
References
123
TEAM LRN
13
Hybrid methods
Zbign ie w Michale w icz
There is some experimental evidence (Davis 1991, Michalewicz 1993) that the
enhancement of evolutionary methods by some additional (problem-specific)
heuristics, domain knowledge, or existing algorithms can result in a system
with outstanding performance. Such enhanced systems are often referred to as
hjhrid e v ol ut i on ary systems.
Several researchers have recognized the potential of such hybridization of
evolutionary systems. Davis ( 199 1, p 56) wrote:
When I talk to the user, I explain that niy plan is to hybridize the
genetic algorithm technique and the current algorithm by employing
the following three principles:
0
[...I I use the term hybrid genetic algorithm for algorithms created by
applying these three principles.
The above three principles emerged as a result of countless experiments of
many researchers, who tried to 'tune' their evolutionary algorithms to some
problem at hand, that is, to create 'the best' algorithm for a particular class
of problems. For example, during the last 15 years, various applicationspecific variations of evolutionary algorithms have been reported (Michalewicz
1996); these variations included variable-length strings (including strings whose
elements were ij4tzeti-else rules), richer structures than binary strings, and
experiments with modified genetic operators to meet the needs of particular
applications. Some researchers (e.g. Grefenstette 1987) experimented with
incorporating problem-specific knowledge into the initialization routine of an
evolutionary system; if a (fast) heuristic algorithm provides individuals of the
124
TEAM LRN
Hybrid methods
125
I26
Hybrid methods
References
Adler D 1993 Genetic algorithms and simulated annealing: a marriage propo5al Proc.
IEEE lnt. Cot$ o t i Neurd Nehtwrks pp 1 104-9
Angeline P J 1995 Morphogenic evolutionary computation: introduction, issues. and
examples Proc. 4th Anti. Cot$ o t i E\*olutiotiury Progrmnming (Sun Diego, CA,
March l9Y5) ed J R McDonnell, R G Reynolds and D B Fogel (Cambridge. M.4:
MIT Press) pp 387-401
Davis L 1991 Handbook of Genetic Al,ip-ithtm (New York: Van Nostrand Reinhold)
Grefenstette J J 1987 Incorporating problem specific knowledge into genetic algorithms
Genetic. Aigorithm utid Siniuiuted Aiitieaiittg ed L Davi5 (Lo\ AI tos, CA: Morgan
Kaufmann) pp 42-60
Michalewicz 2 I993 A hierarchy of evolution programs: an experimental study
E,dictionqt Cotnplct. 1 5 1-76
-1996 Genetic Algorithms + D(itLi Struc-tures = E , d u t i o n Pmgmtm 3rd edn (New
York: Springer)
Miihlenbein H, Gorges-Schleuter M and Kramer 0 1988 Evolution algorithms in
combinatorial optimimtion Purdlel Cotnput. 7 65-85
TEAM LRN
14
Introduction to representations
Kalyanrnoy Deb
127
128
Introduction to representations
I29
Important representations
applications, decision variables are directly used and modified genetic operators
are used to make a successful search. A detailed discussion of the real-valued
vector representations is given in Chapter 16.
In evolution strategy (ES) and evolutionary programming (EP) studies, a
natural representation of the decision variables is used where a real-valued
solution vector is used. The numerical values of the decision variables are
immediately taken from the solution vector to compute the objective function
value. In both ES and EP studies, the crossover and mutation operators are used
variable by variable. Thus, the relative positioning of the decision variables in
the solution vector is not an important matter. However, in recent studies of
ES and EP, in addition to the decision variables, the solution vector includes a
set of strategy parameters specifying the variance of search mutation for each
variable and variable combinations. For n decision variables, both methods use
an additional number between one and n ( n 1)/2 such strategy parameters,
depending on the degree of freedom the user wants to provide for the search
algorithm. These adaptive parameters control the search of each variable,
considering its own allowable variance and covariance with other decision
variables. We discuss these representations in Section 16.2.
In permutation problems, the solutions are usually a vector of node identifiers
representing a permutation. Depending on the problem specification, special care
is taken in creating valid solutions representing a valid permutation. In these
problems, the absolute positioning of the node identifiers is not as important as
the relative positioning of the node identifiers. The representation of permutation
problems is discussed further in Chapter 17.
In early EP works, finite-state machines were used to evolve intelligent
algorithms which were operated on a sequence of symbols so as to produce an
output symbol which would maximize the algorithms performance. Finite-state
representations were used as solutions to the underlying problem. The input
and output symbols were taken from two different finite-state alphabet sets. A
solution is represented by specifying both input and output symbols to each link
connecting the finite states. The finite-state machine tran forms a sequence of
input symbols to a sequence of output symbols. The finite-state representations
are discussed in Chapter 18.
In genetic programming studies, a solution is usually a LISP program
specifying a strategy or an algorithm for solving a particular task. Functions
and terminals are used to create a valid solution. The syntax and structure of
each function are maintained. Thus, if an OR function is used in the solution,
at least two arguments are assigned from the terminal set to make a valid OR
operation. Usually, the depth of nestings used in any solution is restricted to
a specified upper limit. In recent applications of genetic programming, many
special features are used in representing a solution. As the iterations progress,
a part of the solution is frozen and defined as a metafunction with specified
arguments. We shall discuss these features further in Chapter 19.
TEAM LRN of a solution is important in the
As mentioned earlier, the representation
130
Introduction to representations
131
References
References
Deb K 1995 0ptimi:utionfiw Engineering Design: AIgorithnt\ r r n d E.triinples (New Delhi:
Prentice-Hall)
-I997 A robust optimal design technique for mechanical component design
Ewlution(in Algorithnts in Eiigirieeririg Applicutioris ed D Dasgupta and Z
Michalewicz (Berlin: Springer) in press
Deb K and Agrawal R 1995 Simulated binary crossover for continuous search space
C0nrple.u syst. 9 1 15-48
Chaturvedi D, Deb K and Chakrabarty S K 1995 Structural optimization using real-coded
genetic algorithms Proc. Syntp. on Genetic Algorithiiis (Detir(itiirii)cd P K Roj and
S D Mehta (Dehradun: Mahendra Pal Singh) pp 73-82
Eshelman L J and Schaffer J D 1993 Real-coded genetic algorithms and i n t e n d whemata
Founditions of Genetic Algoritliriis I1 ( k i l , CO) ed D Whitley (San Matco. CA:
Morgan Kaufmann) pp 187-202
Kargupta H, Deb K and Goldberg D E 1992 Ordering genetic algorithms and deception
Purullel Problem Solvitig .from Nature I1 (Brussels) ed R Manner and B Manderich
(Amsterdam: North-Holland) pp 47-56
Radcliffe N J I993 Genetic set recombination Fouiidutioits of Getretic Algorithi~isI1 ( Vtril,
CO) ed D Whitley (San Mateo, CA: Morgan Kaufmann) pp 203-19
Reklaitis G V, Ravindran A and Ragsdell K M 1983 Engineering 0ptiiiri:ciiioii: Mcthorls
cind Applicutions (New York: Wiley j
Schaffer J D, Caruana R A, Eshelman L J and Das R 1989 A study of control
parameters affecting online performance of genetic algorithms P r w . 3rd lilt. Car$
011 Genetic. Algorittinis (Fuirfis, WA, 1989) ed J D Schaffer (San Mateo. CA:
Morgan Kaufmannj pp 5 1-60
Wright A 1991 Genetic algorithms for real parameter optimiration Foiriicltitioii.c of
Genetic Algorithms (Bloonzirrgton, IN) ed G J E Rawlins (San Mateo, CA: Morgan
Kaufmann) pp 205-20
TEAM LRN
15
Binary strings
Thornas Back
The classical representation used in so-called canonical genetic algorithms
consists of binary vectors (often called bitstrings or binary strings) of fixed length
t ; that is, the individual space I is given by I = (0, I }' and individuals a E I are
denoted as binary vectors a = (al,. . . , a t ) E (0, I ] ' (see the book by Goldberg
( 1989)). The mutation operator (Section 32.1) then typically manipulates these
vectors by randomly inverting single variables CL, with small probability, and
the crossover operator (Section 33. I ) exchanges segments between two vectors
to form offspring vectors.
This representation is often well suited to problems where potential
solutions have a canonical binary representation, i.e. to so-called pseudo-Boolean
optimization problems of the form f' : (0, I ) ' -+ R. Some examples of such
combinatorial optimization problems are the maximum-independent-set problem
in graphs, the set covering problem, and the knapsack problem, which can be
represented by binary vectors simply by including (excluding) a vertex, \et.
or item i in (from) a candidate solution when the corresponding entry ( I , = 1
( ( I , = 0).
Canonical genetic algorithms, however, also emphasize the binary
representation in the case of problems f' : S -+ R where the search space S
fundamentally differs from the binary vector space (0, I } ' . The most prominent
example of this is given by the application of canonical genetic algorithms for
continuous parameter optimization problems f' : R" + R as outlined by Holland
( 1975) and empirically investigated by De Jong (1975). The mechanism\ of
encoding and decoding between the two different spaces (0, I } ' and R" then
require us to restrict the continuous space to finite intervals [ U , ,U , ] for each
variable s, E R, to divide the binary vector into I I segments of ( i n most
cases) equal length t , , such that t = t i t t , and to interpret a subsegment
(cJ,,-~)~,+I,
. . . , q t ,) (i = I , . . . , U ) as the hinary encoding of the variable
x, . Decoding then either proceeds according to the standard binary decoding
function r' : (0, I}' --+ [ U , .q ] , where (see Back 1996)
(15.1)
I32
TEAM LRN
Binary strings
133
or by using a Gray code interpretation of the binary vectors, which ensures that
adjacent integer values are represented by binary vectors with Hamming distance
one (i.e. they differ by one entry only). For the Gray code, equation ( 1 5. I ) is
extended by a conversion of the Gray code representation to the standard code,
which can be done for example according to
n:=l[~,,
134
Binary strings
are likely to be discrete because they aim at modeling the adaptive capabilities
of natural evolution on the genotype level.
Interpreting a genetic algorithm as an algorithm that processes schemata,
Holland (1975, p 7 I ) then argues that the number of schemata available under
a certain representation is maximized by using binary variables; that is, the
maximum number of schemata is processed by the algorithm if ci, E (0, I ) .
This result can be derived by noticing that, when the cardinality of an alphabet
A for the allele values is k = ( A ( ,the number of different schemata is ( k I ) '
(i.e. 3' in the case of binary variables). For binary alleles, 2' different solutions
can be represented by vectors of length t , and in order to encode the same
number of solutions by a k-ary alphabet, a vector of length
e'= c--In 2
Ink
( I 5.3)
135
References
and obey some structure preserving conditions that still need to be formulated
as a guideline for finding a suitable encoding.
References
Back T 1993 Optimal mutation rates in genetic search Proc. 5th Inr. Cot$ on Genetic
Algorithms (Urbartri-Chatnpaign, IL, July 1993) ed S Forrest (San Mateo, CA:
Morgan Kaufmann) pp 2-8
-1
996 E\dutionun Algorithms in T h e o n und Prwtice (New York: Oxford
University Press)
Bean J C I993 Genetics urtd Randotn Keys for Sequences ~ i n dOptitnixtion Technical
Report 9 2 4 3 , University of Michigan Department of Industrial and Operations
Engineering
Davis L 1991 Handbook of Genetic Algorithtits (New York: Van Nostrand Reinhold)
De Jong K A 1975 An Anulyis cf the Behaviour of a Cl~.s.scf Genetic Aduptiite Swtents
PhD Thesis, University of Michigan
Goldberg D E 1989 Genetic Algorithm in Search, 0ptitni:ution arid MucAine Letrrning
(Reading, MA: Addison Wesley)
-1
99 1 The theory of virtual alphabets Parallel Problent Soliing .front Nuticre-Proc.
1st Workshop, PPSN I (Lecture Notes in Computer Science 496) ed H - P Schwefel
and R Manner (Berlin: Springer) pp 13-22
Holland J H 1975 Aduptation in Natural and Art$ciul Systetns (Ann Arbor. MI:
University of Michigan Press)
Michalewicz Z 1996 Genetic Algorithms
Data Structures = Eidution ProgruniA
(Berlin: Springer)
Nakano R and T Yamada 1991 Conventional genetic algorithms for job shop problems
Proc. 4th Int. Cot$ on Genetic Algorithm (San Diego, CA, Jidy 1991) ed R K Belew
and L B Booker (San Mateo, CA: Morgan Kaufmann) pp 474-9
Yamada T and R Nakano 1992 A genetic algorithm applicable to large-scale job-shop
problems Parullel Problem Solving from Nature, 2 (Proc. 2nd Int. Con$ on Purullt4
Problem Solving from Nature, Brussels, 1992) ed R Manner and B Manderick
(Amsterdam: Elsevier) pp 28 1-90
-+
TEAM LRN
16
~~
~~
~~~~
Real-valued vectors
David B Fogel
When posed with a real-valued function optimization problem of the form find
the vector x such that F ( x ) : R -+ R is minimized (or maximized), evolution
strategies (Back and Schwefel 1993) and evolutionary programming (Fogel
1995, pp 75-84, 136-7) typically operate directly on the real-valued vector
x (with the components of x identified as object parameters). In contrast,
traditional genetic algorithms operate on a coding (often binary) of the vector
x (Goldberg 1989, pp 80-4). The choice to use a separate coding rather than
operating on the parameters themselves relies on the fundamental belief that
i t is useful to operate on subsections of a problem and try to optimize these
subsections (i.e. building blocks) in isolation, and then subsequently recombine
them so as to generate improved solutions. More specifically, Goldberg ( 1989,
p 80) recommends
The user should select a coding so that short, low-order schemata are
relevant to the underlying problem and relatively unrelated to schemata
over other fixed positions.
The user should select the smallest alphabet that permits a natural
expression of the problem.
Although the smallest alphabet generates he greatest implicit parallelism, there
is no empirical evidence to indicate that binary codings allow for greater
effectiveness or efficiency in solving real-valued optimization problems (see
the tutorial by Davis (1991, p 63) for a commentary on the ineffectiveness of
binary codings).
Evolution strategies and evolutionary programming are not generally
concerned with the recombination of building blocks in a solution and do not
consider schema processing. Instead, solutions are viewed in their entirety, and
no attempt is made to decompose whole solutions into subsections and assign
credit to these subsections.
I36
TEAM LRN
137
With the belief that maximizing the number of schemata being processed
is not necessarily useful, or may even be harmful (Fogel and Stayton 1994),
there is no compelling reason in a real-valued optimization problem to act
on anything except the real values of the vector x themselves. Moreover,
there has been a general trend away from binary codings within genetic
algorithm research (see e.g. Davis 199 1, Belew and Booker I99 1, Forrest
1993, and others). Michalewicz (1992, p 82) indicated that for real-valued
numerical optimization problems, floating-point representations outperform
binary representations because they are more consistent and more precise and
lead to faster execution. This trend may reflect a growing rejection of the
building block hypothesis as an explanation for how genetic algorithms act as
optimization procedures.
With evolution strategies and evolutionary programming, the typical method
for searching a real-valued solution space is to add a multivariate zero-mean
Gaussian random variable to each parent involved in the creation of offspring
(see Section 32.2). In consequence, this necessitates the setting of the covariance
matrix for the Gaussian perturbation. If the covariances between parameters
are ignored, only a vector of standard deviations in each dimension is required.
There are a variety of methods for setting these standard deviations. Section 32.2
offers a variety of procedures for mutating real-valued vectors.
I38
Real-valued vectors
References
Back T and Schwefel H-P 1993 An overview of evolutionary algorithms for parameter
optimization E\dietioncint Cornpiet. 1 1-24
Belew R K and Booker L B (eds) 1991 Proc.. 4th Int. Cot$ on Getietic AIgorith~m(Strti
Diego, CA, July I Y Y I ) (San Mateo, CA: Morgari Kaufmann)
Davis L 1991 A genetic algorithms tutorial IV. Hybrid genetic algorithms Hmzdhook of
Genetic Algorithms ed L Davis (New York: Van Nostrand Reinhold)
Fogel D B 1995 Eidictiomin Compictcition: Rnzurd ci New. Philosophy of Machirie
lntrlligrwe (Piscataway, NJ: IEEE)
Fogel D B. Fogel L J and Atmar J W 1991 Meta-evolutionary programming Proc. 25th
A\ilomcir Cot$ o t i Signcils, Systenis, eind Cornpicters ed R R Chen (San Jose, CA:
Maple) pp 540-5
FogeI D B and Stayton L C 1994 On the effectiveness of crossover in 4mulated
evolutionary optirniration BioSystem 32 17 1-82
Forrest S (ed) I993 Proc. 5th liit. Cotif: on Genetic Algorithnis ( Urhcinci-Chnmpuigri,IL,
Jicly 1993) (San Mateo, CA: Morgan Kaufmann)
Goldberg D E 1989 Geizetic Algorithnis iti Search, 0ptitni:ution mid Machine Letirtiirig
(Reading, MA: Addison-Wesley )
Michalewicr Z 1992 Genetic Algorithms + Dcitii Structures = E~~olutiorrProgrtims
(Berlin: Springer)
Ostermeier A, Gawelc~ykA and Hansen N 1994 A derandomized approach to selfadaptation of evolution strategies Evolcctiotiay Cornput. 2 36940
Rechenberg I 1993 Personal communication
Reed J, Toombs R and Barricelli N A 1967 Simulation of biological ebolution and
machine learning J . Tlieor. Biol 17 3 1 9 4 2
Schwefel H-P I98 I Nitnirritul Optinii,-titiono j Corziputsr Models (Chichester: Wiley )
Spears W M I995 Adapting croswver in evolutionary algorithms Etolutioticiry
Progrtimmirig IV: Proc. 4th Anti. Cord. otz E\wlictioncin Prograniniiiig (Stiti Diego.
CA, MLirck 1995) ed J R McDonnell, R G Reynold\ and D B Fogel (Cambridge,
MA: MIT Press) pp 367-84
TEAM LRN
17
Permutations
Darrell Whitley
17.1 Introduction
To quote Knuth (1973), A permutation of a finite set is an arrangement of its
elements into a row. Given n unique objects, n ! permutations of the objects
exist. There are various prcperties of permutations that are relevant to the
manipulation of permutation representations by evolutionary algorithms, both
from a representation point of view and from an analytical perspective.
As researchers began to apply evolutionary algorithms to applications that are
naturally represented as permutations, it became clear that these problems pose
different coding challenges than traditional parameter optimization problems.
First, for some types of problem there are multiple equivalent solutions. When
a permutation is used to represent a cycle, as in the traveling salesman problem
(TSP), then all shifts of the permutation are equivalent solutions. Furthermore,
all reversals of a permutation are also equivalent solutions. Such symmetries
can pose problems for evolutionary algorithms that rely on recombination.
Another problem is that permutation problems cannot be processed using
the same general recombination and mutation operators which are applied to
parameter optimization problems. The use of a permutation representation may
in fact mask very real differences in the underlying combinatorial optimization
problems. An example of these differences is evident in the description of
classic problems such as the TSP and the problem of resource scheduling.
The traveling salesman problem is the problem of visiting each vertex (i.e.
city) in a full connected graph exactly once while minimizing a cost function
defined with respect to the edges between adjacent vertices. In simple terms,
the problem is to minimize the total distance traveled while visiting all the cities
and returning to the point of origin. The TSP is closely related to the problem
of finding a Hamiltonian circuit in an arbitrary graph. The Hamiltonian circuit
is a set of edges that form a cycle which visits every vertex exactly once.
It is relatively easy to show that the problem of finding a set of Boolean
values that yield an evaluation of true for a three-conjunction normal form
Boolean expression is directly polynomial-time reducible to the problem of
TEAM LRN
I39
Permutations
130
= ( a , a, b, b , b, c , d , e , e , f , f )
there are two a's, three b * \ , one c, one d, two e's and two f ' s , and duplicates
are significant. In xheduling applications that map jobs to machines, it may
be necessary to xhedule two jobs of type a, three jobs of type b, and so on.
Note that it is not necessary that all jobs of type a be scheduled contiguously.
While M in the above illu\tration contains 1 1 elements. there are not I 1 ! unique
permutations. Rather, the number of unique permutations is given by
1I!
TEAM LRN
? '._
3!I !I !2!2!
A
.
141
and in general
n!
nl ! n 2 ! n 3 ! .. .
where n is the number of elements in the multiset and n , is the number of
elements of type i (Knuth 1973). Radcliffe (1993) considers the application of
genetic and evolutionary operators when the solution is expressed as a set or
multiset (bag).
Before looking in more detail at the relationship between permutations and
evolutionary algorithms, some general properties of permutations are reviewed
that are both interesting and useful.
al
2
a2
3
a3
...
...
I?
at1
The inverse is obtained by reordering both rows such that the second row is
LRN
transformed into the sequence 123.TEAM
. . n ; the
reordering of the first that occurs as
I42
LIILI~(~~
59 1826473
i23456789
) (
=
where U!,x. y , and z are variables representing the elements of the permutation
(e.g. U' = 3, .v = I , y = 2, 2 = 4). If ul.ryr: now represents the canonically
ordered permutation 1234,
TEAM LRN
143
Matrix representations
We can also relate this mapping operator to the process of finding an inverse.
The permutations in the expression
r3421,1342(3 124) = r1-$32.7143
( 1234)
are included as rows in an array. To map the left-hand side of the preceding
expression to the terms in the right-hand side, first compute the inverses for
each of the terms in the left-hand side:
( 3421
) ( 4312
1234 )
1234
=
( 3124
1234 )
( ii?: ) *
=
Collect the three inverses into a single array. We also then add 1 2 3 4 to the
array and inverse the permutation 2 3 I 4, at the same time rearranging all the
other permutations in the array:
[ 1 (:!i
4312
1423
2314
I234
3124
This yields the permutations 1432, 2143, and 1234 which represent the
desired canonical form as it relates to the notion of substitution into a symbolic
canonical form. One can also reverse the process to find the permutations p ,
and p, in the following context:
I44
Permutations
c
D
E
F
A B C D E F
0 1 0 0 0 1
1 0 1 0 0 0
0 1 0 1 0 0
0 0 1 0 1 0
0 0 0 1 0 1
1 0 0 0 1 0
c
D
E
F
A
0
0
1
0
0
B
0
0
0
0
1
C D E F
1 0 0 1
0 0 1 1
0 1 0 0
1 0 1 0
0 1 0 0
c
D
E
0
#
#
#
O
#
# 0 0 1
# O # #
0 1 0 0
Alternative representations
A
B
c
D
E
F
A
0
0
0
0
0
0
B
1
0
0
0
0
0
C
1
1
0
0
0
0
145
D E
1 1
1 1
1 1
0 1
0 0
0 0
F
1
1
1
1
1
0
A
B
c
D
E
F
A
0
1
1
1
1
1
B
0
0
1
1
1
0
C
0
0
0
0
0
0
D
0
0
1
0
0
0
E
0
0
1
1
0
0
F
0
1
1
1
1
0
In this case, the lower triangle of the matrix flags im*ei-sioiis.which should
not be confused with an iniierse. If u I c i z ( i l . . . ci,, is a permutation of the
canonically ordered set 1 , 2, 3, . . . , 11 then the pair ( ( I , , N , ) is an iiilvrsioii i f
i < j and u, > CI, (Knuth 1973). Thus, the number of I bits in the lower
triangles of the above matrices is also a count of the number of inLrersions
(which should also not be confused with the itilvrsioiz operator used in simple
genetic algorithms, see Holland 1975, p 106, Goldberg 1989, p 166).
The common information can also extracted as before. This produces the
following matrix:
A
A
B
O
#
#
0
# # # #
# # # 1
D
E
F
#
#
#
# 0 0 1 1
# 0 0 0 1
O O O O O
Note that this binary matrix is again symmetric around the diagonal, except
that the lower triangle and upper triangle have complementary bit values. Thus
only N ( N - 1)/2 elements are needed to represent relative order information.
There have been few studies of how recombination crossover operators
generate offspring in this particular representation space. Fox and McMahon
( 199 1 ) offer some work of this kind and also define several operators that work
directly on these binary matrices for relative order.
While matrices may not be the most efficient form of implementation, they
do provide a tool for better understanding sequence recombination operators
designed to exploit relative order. It is clear that adjacency and relative order
relationships are different and are best expressed by different binary matrices.
Likewise, absolute position information also has a different matrix representation
(for example, rows could represent cities and the columns represent positions).
Cycle crossover (Section 33.3.6; see Starkweather et crl 1991, Oliver er 01 1987)
appears to be a good absolute position operator, although it is hard to find
problems in the literature where absolute position is critical.
I46
Permutations
To illustrate:
= U
hcd e j g h
I = 6 2 5 3 8 7 1 4 which represents P = g h d
It
CI
j e.
This may seem like a needless indirection, but consider that I can
be generalized to allow a larger number of possible values than there are
permutation elements. I can also be generalized to allow all real values
(although for computer implementations the distinction is somewhat artificial
since all digital representations of real values are discrete and finite). We
now have a parameter-based presentation of the permutation such that we can
generate random vectors I representing permutations. If the number of values
for which elements in I are defined is dramatically larger than the number of
elements in the permutation, then duplicate values in randomly generated vectors
will occur with very small probability.
This representation allows a permutation problem to be treated as if i t were
a more traditional parameter optimization problem with the constraint that no
two elements of vector I should be equal, or that there is a well defined way
to resolve ties. Evolutionary algorithm techniques normally used for parameter
optimization problems can thus be applied to permutation problems using this
representation.
This idea has been independently invented on a couple of occasions. The
first use of this coding method was by Steve Smith of Thinking Machines. A
version of this coding was used by the ARGOT Strategy (Shaefer 1987) and the
representation was picked up by Syswerda (1989) and by Schaffer er ul (1989)
for the TSP. More recently, a similar idea was introduced by Bean (1994) under
the name random keys.
17.7
Goldberg and Lingle (1985) built on earlier work by Franz (1972) to describe
similarity subsets between different permutations. Franzs calculations were
related to the use of inversion operators for traditional genetic algorithm binary
representations. The use of inversion operators is very much relevant to the
topic of permutations, since in order to apply inversion the binary alleles must be
tagged in some way and inversion acts in the space of all possible permutations
of allele orderings. Thus,
((6 0 ) ( 3 1) ( 2 0 ) ( 8 1) (1 0 ) (5 1) (7 0 ) (4 0 ) )
is equivalent to
TEAM LRN
147
( ( 1 0 ) ( 2 0)(3 1 ) ( 4 0 ) ( 5 1 ) ( 6 0 ) (7 0 ) ( 8 1))
l !
7 3 !
represents all permutations with a one as the third element. a seven as the
sixth element, and a three as the seventh element. Given o selected positions
in a permutation of length I , there are ( I - o)! permutations that match an oschemata. One can also count the number of possible o-schema. There are
clearly (:)) ways to choose o fixed positions; there are also (,:) ways to pick the
permutation elements that fill the slots, and o ! ways of ordering the elements
(i.e. the number of permutations over the chosen combination of subelements).
Thus, Goldberg (1989, Goldberg and Lingle 1985) notes that the total number
of o-schemata, no,, can be calculated by
Note that in this definition of the o-schemata, relative order is not accounted for.
In other words, if relative order is important then all of the following shifted
o-schemata,
1
!
!
1
!
!
7
!
!
l
3
7
!
3
!
!
!
!
7
!
!
!
3
7
!
3
could be viewed as equivalent. Such schemata may or may not wrap around.
Goldberg discusses o-schemata which have an absolute fixed position (oschemata, type a) and those with relative position which are shifts of il specified
template (0-schemata, type r).
This work on o-schemata predates the distinctions between relative
order permutation problems, absolute position problems, and adjacency-based
problems. Thus, o-schemata appear to be better for understanding resource
scheduling applications than for the TSP. In subsequent work, Kargupta et
nl (1992) attempt to use ordering schemata to construct deceptive functions
for ordering problems-that is, problems where the average fitness values of
the o-schemata provide misleading information. Note that such problems are
constructed to mislead simple genetic algorithms and may or may not be difficult
TEAM LRN(For a discussion of deception see
with respect to other types of algorithm.
148
Permutations
the article by Goldberg ( 1987) and Whitley (1991) and for another perspective
see the article by Grefenstette (1993).) The analysis of Kargupta et ul ( 1 992)
considers PMX. a uniform ordering crossover operator (UOX), and a relative
ordering crossover operator (ROX).
An alternative way of constructing relative order problems and of
comparing the similarity of permutations is given by Whitley and Yoo
(1995). Recall that a relative order matrix has a I bit in position ( X , Y ) if
row element X appears before column element Y in a permutation. Note
that the matrix representation yields a unique binary representation for each
permutation. Using this representation one can also define the Hamming
distance between two permutations PI and P2; Hamming distance is denoted
by HD(index(P1), index(P2)), where the permutations are represented by their
integer index. In the following examples, the Hamming distance is computed
with respect to the lower triangle (i.e. it is a count of the number of 1 bits in
the lower triangle):
A B C D
-- - - - - ---
A B C D
A
B
C
D
I 0 1 1 1
l O O l l
l O O O l
HD(0,O) = 0
I 0 0 0 0
A B C D
-- ---- - --
B D C A
A I 0 0 0 0
B I l O l l
cl
HD(0,ll)
HD(0,23)
1 0 0 0
D 1 1 0 1 0
A B C D
- -- - -- - -D C B A
A
B
C
D
I
J
l
l
O
l
l
l
O
0
l
l
O O
0 0
O O
l O
Whitley and Yoo (1995) point out that this representation is not perfect.
Since 2"') > N ! , certain binary strings are undefined. For example, consider
the following upper triangle:
1 1 1
0 1
0
but
149
4 > 3
References
Bean J C 1994 Genetic algorithms and random keys for sequencing and optimization
ORSA J. Cornpiit. 6
Cormen T, Leiserson C and Rivest R 1990 Introduction to Algorithnis (Cambridge. MA:
MIT Press)
Deb K and Goldberg D 1993 Analyzing deception in trap functions Fouririntions of
Genetic Algorithms 2 ed D Whitley (San Mateo, CA: Morgan Kaufmann)
Fox B R and McMahon M B 1991 Genetic Operrrtors .for Seyirencing Prohleim
Foiindcctiorts of Genetic Algorithrns ed G J E Rawlins (San Mateo, CA: Morgan
TEAM LRN
Kaufmann) pp 284-300
150
Permutations
TEAM LRN
18
Finite-state representations
David B Fogel
18.1 Introduction
A finite-state machine is a mathematical logic. It is essentially a computer
program: it represents a sequence of instructions to be executed, each depending
on a current state of the machine and the current stimulus. More formally, a
finite-state machine is a 5-tuple
A4 = ( Q , t,p , S ,
O)
where Q is a finite set, the set of states, t is a finite set, the set of input symbols,
p is a finite set, the set of output symbols, s : Q x t --+ Q is the next state
function, and o : Q x t + p is the next output function.
Any 5-tuple of sets and functions satisfying this definition is to be interpreted
as the mathematical description of a machine that, if given an input symbol .r
while it is in state q , will output the symbol o ( y , .r) and transition to state s ( q . x).
Only the information contained in the current state describes the behavior of the
machine for a given stimulus. The entire set of states serves as the 'memory' of
the machine. Thus a finite-state machine is a transducer that can be stimulated
by a finite alphabet of input symbols, that can respond in a finite alphabet
of output symbols, and that possesses some finite number of different internal
states. The corresponding input-output symbol pairs and next-state transitions
for each input symbol, taken over every state, specify the behavior of any finitestate machine, given any starting state. For example, a three-state machine is
shown in figure 18.1. The alphabet of input symbols are elements of the set
{0, I ] , whereas the alphabet of output symbols are elements of the set {a,/3, y ]
(input symbols are shown to the left of the slash, output symbols are shown to
the right). The finite-state machine transforms a sequence of input symbols into
a sequence of output symbols. Table 18.1 indicates the response of the machine
to a given string of symbols, presuming that the machine is found in state C.
It is presumed that the machine acts when each input symbol is perceived and
the output takes place before the next input symbol arrives.
TEAM LRN
151
152
Finite-state representations
Figure 18.1. A three-state finite machine. Input symbols are shown to the left of the
slash. Output symbols are to the right of the slash. Unless otherwise specified, the
machine is presumed to start in state A . (After Fogel et nl 1966, p 12).
Table 18.1. The response of the tinite-state machine shown in figure 18.1 to a string of
symbols. In this example, the machine starts in state C .
Present state
Input symbol
Next state
Output symbol
C
B
B
1
C
a!
C
1
A
y
A
0
B
6 6
a!
I
A
18.2 Applications
Finite-state representations are often convenient when the required solutions to
a particular problem of interest require the generation of a sequence of symbols
having specific meaning. For example, consider the problem offered by Fogel
et cil (1966) of predicting the next symbol in a sequence of symbols taken from
some alphabet A (here, T = p = A ) . A population of finite-state machines is
exposed to the environment, that is, the sequence of symbols that have been
observed u p to the current time. For each parent machine, as each input symbol
is offered to the machine, each output symbol is compared with the next input
symbol. The worth of this prediction is then measured with respect to the
given payoff function (e.g. all-none, absolute error, squared error, or any other
expression of the meaning of the symbols). After the last prediction is made,
a function of the payoff for each symbol (e.g. average payoff per symbol)
indicates the fitness of the machine. Offspring machines are created through
mutation (Section 32.4) and/or recombination (Section 33.4). The machines that
provide the greatest payoff are retained to become parents of the next generation.
TEAM LRN
This process is iterated until an actual
prediction of the next symbol (as yet
153
Applications
C,DID
D,CIC
Legend
I
Stan State
,=
C = Cooperate
D = Delecl
Figure 18.2.
A finite-state machine evolved in prisoners dilemma experiThe input symbols form the set
ments detailed by Fogel (1995b, p 215).
((C. C). (C, D),( D ,
C), ( D ,
D)}and the output symbols form the set (C,D ] . The
machine also has an associated first move indicated by the arrow; here the machine
cooperates initially then proceeds into state 6.
IS4
Finite-state representations
References
Angeline P J and Pollack J B 1993 Evolutionary module acqui4ition Proc. 2ndAtiti. Cont.
o t i Eidictioritirv Progr~iiriiriirig( S m Diego,CA) ed D B Fogel and W Atniar (La
Jolla, CA: Evolutionary Prograniming Society) pp I 54-63
Fogel D B 1991 The evolution of intelligent deci\ion-making in gaming Cvber.net. Sv\t.
22 223-36
-1993
E v o l ~ing behavior5 in the iterated prisoner-5 dilemtna E\*olirt.Cornput. 1 77-97
-1995a
On the relationship between the duration o f an encounter and the ebvlution
o f cooperation in the iterated prisoner's dilemma E\*olitt. Coriipict. 3 349-63
__ 1995b E\*olutiotiIir?~
Coinpirtutioti*Touurd II N o t . Philosophy of Mtrchine ltitel1igenc.e
(Piwataway. NJ: IEEE)
Fogel L J, Owens A J and Walsh M J 1966 Artijkicil Ititelligetic~e TIiroirgh Sitiiirlcittd
E\wliitioti (New York: Wiley )
Jeffttrson D, Collins R, Cooper C, Dyer M, Flowers M, Korf R, Taylor C and Wang A
1991 Evolution as a theme i n artificial life: the Genesysflracker \ystcm Artifir*rcrl
Lde I1 ed C G Langton, C Taylor, J D Farmcr and S Ra\mu\wn (Reading, MA:
Addiwn-Wesley ) pp 549-77
TEAM LRN
19
Parse trees
Peter J Angeline
When an executable structure such as a program or a function is the object of
an evolutionary computation, representation plays a crucial role in determining
the ultimate success of the system. If a traditional, syntax-laden programming
language is chosen to represent the evolving programs, then manipulation by
simple evolutionary operators will most likely produce syntactically invalid
offspring. A more beneficial approach is to design the representation to ensure
that only syntactically correct programs are created. This reduces the ultimate
size of the search space considerably. One method for ensuring syntactic
correctness of generated programs is to evolve the desired programs parse
tree rather than an actual, unparsed, syntax-laden program. Use of the parse
tree representation completely removes the syntactic sugar introduced into
a programming language to ensure human readability and remove parsing
ambiguity.
Cramer (1985), in the first use of a parse tree representation in a
genetic algorithm, described two distinct representations for evolving sequential
computer programs based on a simple algorithmic language and emphasized the
need for offspring programs to remain syntactically correct after manipulation
by the genetic operators. To accomplish this, Cramer investigated two encodings
of the language into fixed-length integer representations.
Cramer (1985) first represented a program as an ordered collection of
statements. Each statement consisted of N integers; the first integer identified
the command to be executed and the remainder specified the arguments
to the command. If the command required fewer than N - I arguments.
then the trailing integers in the statement were ignored. Depending on the
syntax of the statements command, an integer argument could identify a
variable to be manipulated or a statement to be executed. Consequently,
even though the program was stored as a sequence it implicitly encoded an
execution tree that could be reconstructed by replacing all arguments referring
to program statements with the actual statement. Cramer (1985) noted that
this representation was not suitable for manipulation by genetic operators and
occasionally resulted in infinite loops when two auxiliary statements referred to
each other.
TEAM LRN
I ss
156
Parse trees
Parse trees
157
evaluated within an implied repeat until done loop that reexecutes the evolved
function until some predetermined stopping criterion is satisfied. For instance,
Koza ( I 992) describes evolving a controller for an artificial ant for which the
fitness function repeatedly applies its program until a total of 400 commands
are executed or the ant completes the task. Numerous examples of such implied
loops can be found in the genetic programming literature (e.g. Koza 1992,
pp 147, 329, 346, Teller 1994, Reynolds 1994, Kinnear 1993).
Often it is necessary to include constants in the primitive language, especially
when mathematical expressions are being evolved. The general practice is to
include as a potential terminal of the language a special symbol that denotes a
constant. When a new individual is created and this symbol is selected to be a
terminal, rather than enter the symbol into the parse tree, a numerical constant
is inserted drawn uniformly from a user-defined range (Koza 1992). Figure 19.1
shows a number of numerical constants that would be inserted into the parse
tree in this manner.
,-,
if - I t - 0
0
0
/\
0.1467
0.547
sin
0.9765
/\
sln
dl
dO
,, ,
/\
, * ,
cis
d3
d2
dl
if - I t - 0
Sin
1.075
dO
Figure 19.1. An example parse tree representation for a complex numerical function. The
function if-It-0 is a numerical conditional that returns the value of its second argument
if its first argument evaluates to a negative number and otherwise returns the value of
its third argument. The function 3- denotes a protected division operator that returns a
value of 1.0 if the second argument (the denominator) is zero.
158
Parse trees
References
Angeline P J 1996 Genetic programmings continued evolution Acl\wices in Genetic
Progrrttrrmirig vol 2, ed P J Angeline and K Kinnear (Cambridge, MA: MIT Prc\\)
pp 1-20
Angeline P J and Pollack J B 1994 Co-evolving high-level representations Artficiul Lije
I l l ed C G Langton (Reading, MA: Addiwn-Wesley) pp 55-71
Cramer N L 1985 A rcprewmtion for the adaptive generation o f 5imple wquential
programs Proc. 1s t lrit. Conj: on Genetic Algorithms (Pittsburgh, PA, Jitly 1985) ed
J J Grefenstette (Hillsdale, NJ: Erlbaum) pp 183-7
Kinnear K E 1993 Generality and difficulty in genetic programming: evolving a sort
Proc. 5th Int. Car$ on Genetic Algorithms ( Utbutiu-Chuinpuigti, IL, Jiily 1993) cd
S Forre\t (San Mateo, CA: Morgan Kaufmann) pp 287-94
Kola J R I992 Genetic Progrurriniing: o r i the Progriintnting o j Conipiiter\ bj*Meuti 5 of
Ntrturctl Selection (Cambridge, MA: MIT Pre\\)
-1994 Get1e tic Prog rutnui irig 11: A ut omutic Disco \Yr\ o j Re us LI ble ProCSrmis
(Cambridge, MA: MIT Press)
Kola J R and Andre D 1996 Classifying protein segments as transmembrane domains
wing architecture-altering operation5 in genetic programming Ad\wzc.e,\ in Genetic
Progrunzniing vol 2, ed P J Angeline and K Kinnear (Cambridge, MA: MIT Pre55)
pp 155-76
TEAM
LRN
programming
E\wliitionuT Compirt. 3 199-230
Montana D J 1995 Strongly typed genetic
References
159
Reynolds C W 1994 Evolution of obstacle avoidance beha\ ior: using noi\e to promote
robust solutions Adilcirrces it? Gewtic Algorithnis ed K Kinnear (Cambridge, MA:
MIT Press) pp 22 111.3
Rosca J P and Ballard D H 1996 Discovery of subroutines in genetic programming
Adwiwes in Gerzetic Progruntniing vol 2, ed P J Angeline and K Kinnear
(Cambridge, MA: MIT Press) pp 177-202
Teller A 1994 The evolution of mental models Adiwic.c),\ in Gcwetic*Algorithms ed
K Kinnear (Cambridge, MA: MIT Press) pp 199-220
TEAM LRN
20
Guidelines for a suitable encoding
David B Fogel and Peter J Angeline
In any evolutionary computation application to an optimization problem, the
human operator determines at least four aspects of the approach: representation,
variation operators, method of selection, and objective function. It could be
argued that the most crucial of these four is the objective function because it
defines the purpose of the operator in quantitative terms. Improperly specifying
the objective function can lead to generating the right answer to the wrong
problem. However it should be clear that the selections made for each of
these four aspects depend in part on the choices made for all the others. For
example, the objective function cannot be specified in the absence of a problem
representation. The choice for appropriate representation, however, cannot
be made in the absence of anticipating the variation operators, the selection
function, and the mathematical formulation of the problem to be solved. Thus,
an iterative procedure for adjusting the representation and search and selection
procedures in light of a specified objective function becomes necessary in many
applications of evolutionary computation. This section focuses on selecting the
representation for a problem, but it is important to remain cognizant of the
interdependent nature of these operations within any evolutionary computation.
There have been proposals that the most suitable encoding for any problem
is a binary encoding (Chapter IS) because it maximizes the number of
schemata being searched implicitly (Holland 1975, Goldberg 1989), but there
have been many examples in the evolutionary computation literature where
alternative representations have provided for algorithms with greater efficiency
and optimization effectiveness when compared with identical problems (see
e.g. the articles by Biick and Schwefel (1993) and Fogel and Stayton (1994)
among others). Davis ( 1991) and Michalewicz (1996) comment that in many
applications real-valued (Chapter 16) or other representations may be chosen to
advantage over binary encodings. There does not appear to be any general
benefit to maximizing implicit parallelism in evolutionary algorithms, and,
therefore, forcing problems to fit binary representation is not recommended.
The close relationship between representation and other facets of
evolutionary computation suggests that, in many cases, the appropriate choice of
representation arises from the operator's ability to visualize the dynamics of the
I60
TEAM LRN
161
f ( x , y ) = x-
+ J3
x,?'
R.
Figure 20.1. A quadratic bowl in two dimensions. The shape of the response surface
suggests a natural approach for optimization. The intuitive choice is to use real-valued
encodings and continuous variation operators. The shape of a response surface can be
useful in suggesting choices for suitable encodings.
I62
References
Back T and Schwefel H-P I993 An overview of evolutionary algorithms for parameter
optiinimtion E \ d i i t i o i i r i n Cornpiit. 1 1-24
Davis L ( e d ) 199 1 H m r l h o o k oj'Cerirtic A1,qorithni.s (New York: Van Nostrand Reinh1.M)
Fogel D B and Stayton L C I994 On the eft'ec(ibmcss of crossover in simulated
evolutionary o p t i m i h o n BioSystenis 32 I 7 1-82
Goldberg D E 1989 C o w t i c AIgorithi)is i i i S r ~ i r ~ Optimiyitiori,
~h.
triitl Metchine Lectriiirix
(Reading, MA: Addison-Wesley
Goldberg D E and Smith R E I987 Nonstationary function optimization using gewtic
algortihms with dominance and tiiploidy Proc. 2nd lut. Cot?$ o i i Gcwetic AI~qorit/irti,s
(Ctimhridge, MA, 1987) ed J J Grefenstette (Hillsdale, NJ: Erlhaum) pp 59-68
Holland J H 1975 Aclciptcitiori in Nlitiirul rind Artijic-icil Systeim (Atin Arhor, MI:
University of Michigan Press)
Michalewicz Z 1996 Genetic Algorithms + Dutct Striwtirres = Elwliitioii Pro,qrrinis 3rd
edn (Berlin: Springer)
Ng K P and Wong K C 1995 A new diploid scheme and dominance change mecha~rixm
for non-stationary function optimization Proc. 6th lnt. Car$ o i i Grnetic Algoritlrriis
(Pittshiirgh, PA. Jirly jYY.5) ed L J Eshelrnan (San Mateo, CA: Morgan Kaufm.inn)
pp 159-66
Schraudolph N N and Beleu R K 1992 Dynamic parameter encoding for geiietic
algorithms Mtrchiiie Ltwrriiri,q 9 9-2 1
TEAM LRN
21
~~
0ther representations
Peter J Angeline and David B Fogel
1M , d
N , Z, 2 N ]
21.2 Introns
In contrast to the above hybridization of different forms of representation,
another nontraditional approach has involved the inclusion of noncoding
regions (introns) within a solution (see e.g. Levenick 1991, Golden et cil 1995.
Wu and Lindsay 1995). Solutions are represented in the form
S I)intronl.rz(intronI. . .
IintronI.r,,
163
1 64
Other representations
References
Angeline P J, Saunders G M and Pollack J B 1994 An evolutionary algorithm that
TEAM
LRNTr(it1.r.Neilrd Netbtwrkr NN-5 54-65
constructs recurrent neural networks
IEEE
I65
References
Back T and Schiitz M 1995 Evolution strategies for mixed-integer optimization of optical
multilayer sy\tem\ Proc. 4th Anit. Coi$ on E,dirtioiirii?) Pmgrcinimiiig (Strri Dic>go,
CA, Murcli 1995) ed J R McDonnell, R G Reynolds and D B Fogel (Cambridge.
MA: MIT Press) pp 33-51
Bagley J D I967 The Behaiior of Adaptiipe Systents \ivhicA Employ Cerietic tirid
Correldon Algorithtns Doctoral Dissertation, University of Michigan: Unik er\ity
Micro fi 1ms 68-7556
Brindle A 198 I Geitc4c Algorithms f o r Fuizctiort 0ptinti;titioir Doctoral Di\sertation,
University of Alberta
Cobb H G and Grefenstette J J 1993 Genetic algorithms for tracking changing
environments Proc. 5th liit. Corzf: on Gerietic Algorithm ( Urb~iirci-CIiiim~~ni~~~i.
IL,
July 1993) ed S Forrest (San Mateo, CA: Morgan Kaufmann) pp 523-30
of Mtdiiric.
Fogel D B 1995 Eidiitiortcir~ Coniputution: T o t i ~ ~ (tI l Nc.\i. Pliilosopli~~
Iiirelligem-e (Piscataway, NJ: IEEE)
Goldberg D E, Korb D E and Deb K 1989 Messy genetic algorithms: motivation, analysis,
and first results Comp1e.rSyst. 3 493-530
Goldberg D E and Smith R E 1987 Nonstationary function optimization using genetic
algorithm\ with dominance and diploidy Proc. 2nd Itit. Car$ o t i Genetic Algorithm
(Ccimbridge, MA, July 1987) ed J J Grefenstette (Hillsdale, NJ: Erlbaum) pp 59-68
Golden J B, Garcia E and Tibbetts C 1995 Evolutionary optimiyation o f a neural netuorkba\ed signal processor for photometric data from an automated DNA sequencer
Proc.. 4th A m . Car$ on E,vlutionan Prograriiriiirig ( S a n Diego, CA, Mlirdi 199.5)
ed J R McDonnell, R G Reynolds and D B Fogel (Cambridge. MA: MIT Press)
pp 579-601
Haffner S B and Sebald A V 1993 Computer-aided design of fuzzy HVAC
controllers using evolutionary programming Proc. 2nd A i i r i . Coi$ o i i Eiwlictioriar-Jq
Prognrmi?iiiig ( S a n Diego, CA, 1993) ed D B Fogel and W Atmar (La Jolla, CA:
Evolutionary Programming Society) pp 98-1 07
Hollstein R B I97 1 Artijcicil Genetic Adaptation in Conipirter Control Systeim Doctoral
Dissertation, University of Michigan; University Microfilms 7 1-23, 773
Levenick J R 1991 Inserting introns improves genetic algorithm wccess rate: taking
a cue from biology Proc. 4th Int. Cotzf: on Genetic Alpwithim ( S m Diego, CA,
Jirly 1991) ed R K Belew and L B Booker (San Mateo, CA: Morgan Kaufmann)
pp 123-27
McDonnell J R and Waagen D 1994 Evolving recurrent perceptrons for time-series
modeling IEEE Traits. Neitrctl N e h w k s NN-5 24-38
Ng K P and Wong K C 1995 A new diploid scheme and dominance change mechanism
for non-stationary function optimization Proc. 6th I t i t . Coi!f: o i i Genetic Algorithm
(Pittsburgh, PA, J u l y 1995) ed L J Eshelman (San Mateo, CA: Morgan Kaufmann)
pp 159-66
Wu A S and Lindsay R K 1995 Empirical studies of the genetic algorithm with noncoding
segments E~dutioiiunComput. 3 I2 1-48
TEAM LRN
22
Introduction to selection
Kalyanmoy Deb
22.1
Working mechanisms
TEAM LRN
167
Pseudocode
22.2 Pseudocode
Some EC algorithms (specifically, genetic algorithms (GAS) and genetic
prograininiiig (GP)) usually apply the selection operator first to select good
solutions and then apply the recombination and mutation operators on these
good solutions to create a hopefully better set of solutions. Other EC algorithms
(specifically, evolution strategies (ES) and adiitioiiury progrcinirnirzg (EP))
prefer using the recombination and mutation operator first to create a set of
solutions and then use the selection operator to choose a good set of solutions.
The selection operator in ( p A ) ES and EP techniques chooses the offspring
solutions from a combined population of parent solutions and solutions obtained
after recombination and mutation. In the case of EP, this is done statistically.
However, the selection operator in ( p ,A ) ES chooses the offspring solutions
only from the solutions obtained after the recombination and mutation operators.
Since the selection operators are different in different EC studies, it is difficult
to present a common code for all selection operators. However, the following
pseudocode is a generic for most of the selection operators used in EC studies.
The parameters U
, and A are the numbers of parent solutions and offspring
solutions after recombination and mutation operators, respectively.
The
parameter q is a parameter related to the operators selective pressure, a matter
we discuss later in this section. The population at iteration t is denoted by
P ( t ) = { a , az,
, . . .} and the population obtained after the recombination and
mutation operators is denoted by f ( t ) = {a,
, a>,. . .}. Since GAS and GP
techniques use the selection operator first, the population P ( t ) before the
selection operation is an empty set, with no solutions. The fitness function
is represented by F ( t ) .
Input: p , A, q , P ( t ) E Z, P ( t ) E Z E b , F ( t )
Output: P ( t ) = {a,,
ay...
I to
for i
a:(t)t S\election(P(t),
return({a,(r),. . . , a;:(t)
E ZL
,U
168
Introduction to selectiol
169
Pseudocode
170
Introduction to selection
ei) chosen from the pool. The complete pool is then sorted in descending order
of this score and the first p solutions are chosen deterministically. Thus, this
selection scheme is similar to the ( p + p ) ES selection scheme with a tournament
selection of q tournament size. Back et cil ( 1994) analyzed this selection scheme
as a combination of ( p p ) ES and tournament selection schemes, and found
some convergence characteristics of this operator.
Goldberg and Deb (1991 ) have compared a number of popular selection
schemes in terms of their convergence properties, selective pressure, takeover
times, and growth factors, all of which are important in the understanding of
the power of different selection schemes used in GA and GP studies. Similar
studies have also been performed by Back et er1 (1994) for selection schemes
used in ES and EP studies. A detailed discussion of some analytical as well as
experimental comparisons of selection schemes is presented in Chapter 29. In
the following section, we briefly discuss the theory of selective pressure and its
importance in choosing a suitable selection operator for a particular application.
takeover time is defined as the speed at which the best solution in the initial
population would occupy the complete population by repeated application of the
selection operator alone (Back 1994. Goldberg and Deb 1991). If the takeover
time of a selection operator is lctrge (that is, the operator takes a large number
of iterations for the best solution to take over the population), the selective
pressure of the operator is smcill, and vice versa. Thus, the selective pressure or
the takeover time is an important parameter for successful operation of an EC
algorithm (Back 1994, Goldberg et crl 1993). This parameter gives an idea of
how greedy the selection operator is in terms of making the population uniform
with one particular solution. If a selection operator has a large selective pressure,
the population loses diversity in the population quickly. Thus. in order to awid
premature convergence to a wrong solution, either a large population is required
or highly disruptive recombination and mutation operators are needed. However.
a selection operator with a small selection pressure makes a slow convergence
and permits the recombination and mutation operators enough iterations to
properly search the space. Goldberg and Deb ( 1991) have calculated takeover
times of a number of selection operators used i n GAS and GP studies and Back
( 1994) has calculated the takeover time for a number of selection operators used
in ES, EP, and GA studies. The former study has also introduced two other
parameters-early and late growth rate-characterizing the selection operators.
The growth rate is defined as the ratio of the number of the best solutions
in two consecutive iterations. Since most selection operators have different
growth rates as the iterations progress, two different growth rates-early and
TEAM
LRNgrowth rate is calculated initially,
late growth rates-are defined. The
early
171
References
when the proportion of the best solution in the population is negligible. The
late growth rate is calculated later, when the proportion of the best solution in
the population is large (about 0.5). The early growth rate is important, especially
if a quick near-optimizer algorithm is desired, whereas the late growth rate can
be a useful measure if precision in the final solution is important. Goldberg
and Deb (1991) have calculated these growth rates for a number of selection
operators used in GAS. A comparison of different selection schemes based on
some of the above criteria is given in Chapter 29.
The above discussion suggests that, for a successful EC simulation. the
required selection pressure of a selection operator depends on the recombination
and mutation operators used. A selection scheme with a large selection pressure
can be used, but only with highly disruptive recombination and mutation
operators. Goldberg et cil ( 1993) and later Thierens and Goldberg ( 1993) ha\ie
found functional relationships between the selective pressure and the probabilit)?
of crossover for successful working of selectorecombinative GAS. These studies
show that a large selection pressure can be used but only with a large probability
of crossover. However. if a reasonable selection pressure is used, GAS ~ . ' o r k
successfully for a wide variety of crossover probablities. Similar studies can
also be performed with ES and EP algorithms.
References
Back T I994 Selectitre pressure in etrolutionary algorithms: a charactcrimtion of selection
mechanisms Proc. 1st IEEE C o i ~ oti
; E~dirtioti~ii;~~
Cornpirtrrtiori ( Orlmrio. Fl,.
1994) (Piscataway, NJ: IEEE) pp 57-62
Back T, Rudolph G and Schwefel H-P 1994 Evolutionary programming and ctwlution
strategies: similarities and differences Proc. 2ritl Aiiri. Col$ o i i ,*olirtioiitir;\~
Progrrinit?iitig ( S r i t i Diego, CA, Jirl?, 1994) ed D B Fogel and W Atmar ( L a Jolla.
CA : E v o 1U t i on ary Program m i ng Soc i et y )
Goldberg D E 1989 Genetic AIgor-ithnis in Setrr-(4, Optii~iixitiou,tirid Mticliirir Lctrr.riiricq
(Reading, MA: Addison-Wesley)
Goldberg D E and Deb K 1991 A comparison of selection schemes used i n genetic
I N ) ed G J E Raw'lins
algorithms Foiinckitions of Genetic Algorithnis ~Bl(~oiiii~i~stoii,
(San Mateo, CA: Morgan Kaufmann) pp 69-93
Goldberg D E, Deb K and Theirens D 1993 Toward a better understanding o f mixing i n
genetic algorithms J. SIC 32 10-6
Thierens D and Goldberg D E 1993 Mixing in genetic algorithms Proc. -5th / l i t . Car$ or1
Genetic. Algorithrm (Ur~?(i,i,i-Chai?iptiigii,
lL, Ji4!\' 1993) ed S Forrest (San Mateo.
CA: Morgan Kaufmann) pp 38-45
TEAM LRN
23
Proportional selection and sampling
algorithms
J o h I i G refe I i stet te
23.1
Introduction
23.2
Fitness functions
172
TEAM LRN
173
Fitness functions
where A , is the object variable space. The objective function typically measures
some cost to be minimized or some reward to be maximized. The definition of
the objective function is, of course, application dependent. The characterization
of how well evolutionary algorithms perform on different classes of objective
functions is a topic of continuing research. However, a few general design
principles are clear when using an evolutionary algorithm.
(i) The objective function must reflect the relevant measures to be optimized.
Evolutionary algorithms are notoriously opportunistic, and there are several
known instances of an algorithm optimizing the stated objectii e function.
only to have the user realize that the objective function did not actually
represent the intended measure.
(ii) The objective function should exhibit some regularities over the space
defined by the selected representation.
(iii) The objective function should provide enough information to dri\fe the
selective pressure of the evolutionary algorithm. For example, needle-ina-haystack functions, i.e. functions that assign nearly equal calue to every
candidate solution except the optimum, should be avoided.
The jitness fiiriction
@:A,+IR+
maps the raw scores of the objective function to a non-negative interval. The
fitness function is often a composition of the objecti\ye function and a scaling
function g:
W 4 ( t ) )= g ( . f ( l N ) ) )
where cc,(t) E A , . Such a mapping is necessary if the goal is to minimize
the objective function, since higher fitness values correspond to lower objectiLre
values in this case. For example, one fitness function that might be used when
the goal is to minimize the objective function is
@ ( a m )= .finax - . f ( G ( t ) )
where fmax is the maximum value of the objective function. If the global
maximum value of the objective function is unknown, an alternatiLre is
@(U&))
= .fmax(t)- . f ( W ) )
1 f ( W )) .fi,,l(t)
where & ( t ) is the minimum observed value of the objective function up to
time t . For maximization problems, this becomes
@ ( a m )=
1 .fmax(t)- f ( % ( t ) )
TEAM LRN
Note that the latter two fitness functions
yield a range of (0, 1 1 .
174
23.2. I
Fitrirss sculirig
= q f ( l l ; ( t ) )-- B ( t )
where 6 is an update rate of, say, 0.1, and fMOr,,(t)is the worst objective value
in the population at time t .
Sigizitr w r l i r i g (Goldberg 1989) is based on the distribution of objective
kralues within the current population. It is defined as follows:
where ~ f ( tis) the mean objective value of the current population, a,.(r)is the
(sample) standard deviation of the objective values in the current population,
and is a constant, say c. = 2. The idea is that f ( t )-cat (I)
represents the Icast
acceptable objective value for any reproducing individual. As the population
improves, this statistic tracks the improvement, yielding a level of seleclive
pressure that is sensitive to the spread of performance values in the population.
Fitness scaling methods based on power laws have also been proposed A
fixed transformation of the form
c s
I75
Selection probabilities
where the parameter T can be used to control the level of selective pressure
during the course of the evolution. It is suggested by de la Maza and Tidor
(1993) that, if T decreases with time as in a simulated annealing procedure,
then a higher level of selective pressure results than with proportional selection
without fitness scaling.
23.4 Sampling
In an incremental, or steady-state, algorithm. the probability distribution can
be used to select one parent at a time. This procedure is commonly called
the roirlette \ t h e 1 sampling algorithm, since one can think of the probability
distribution as defining a roulette wheel on which each slice has a width
corresponding to the individual's selection probability, and the sampling can
be envisioned as spinning the roulette wheel and testing which slice ends up at
the top. The pseudocode for this is shown below:
TEAM LRN
176
i):
(f
Note that the pseudocode allows for any number h > 0 of children to
be \pecitied. If h = I . SUS behaves like the roulette wheel function. For
generational algorithm\, SUS is usually invoked with h = p .
In can he shown that the expected number of offspring that SUS assign\ to
indikidual i i \ A Pr(i), and that on each invocation of the procedure, SUS assigns
either LA Pr(i)J or [A Pr(i)l offspring to individual i. Finally, SUS is optimdly
efficient, making a \in& pa\\ over the individual3 to assign all off9pring.
23.5 Theory
The section presents some results from the theory of proportional selection.
TEAMfollowing
LRN
First. the schema theorem is described,
by a discussion of the effects
I77
Theory
where Q, is the fitness function and & ( t ) denotes the average titness of the
individuals in P ( t ) . The most important feature of proportional selection is
that it induces the following target sampling rates for all hyperplanes i n the
population :
rti(
tsr(H. t ) =
H.t)
tsr(ir,, t )
(23.1)
where @ ( H ,t ) is simply the average fitness of the representatiies of H i n P ( t ) .
This result is the heart of the schema theorem (Holland 1975), which has been
TEAM LRN
called the jiulckri?iei~rcIltlzeoreni of genetic
u l g o r i t h s (Goldberg 1 989 ).
178
Theory
179
Even though both genetic algorithms behave according to the schema theorem.
they clearly allocate trials to hyperplane H at different rates, and thus produce
entirely different sequences of populations. The relationship between the schema
theorem and the objective function becomes even more complex if the fitness
function Q> is dynamically scaled during the course of the algorithm. Clearly.
the allocation of trials described by schema theorem depends on the precise
form of the fitness function used in the evolutionary algorithm. And of course,
crossover and mutation will also interact with selection.
where cri(t) is the fitness variance of the population at time t . From this
formula, it is easy to see that, without dynamic fitness scaling, an evolutionary
algorithm tends to stagnate over time since cri(t) tends to decrease and & ( t )
tends to increase. The fitness scaling techniques described above are intended
to mitigate this effect. In addition, operators which produce random variation
(e.g. mutation) can also be used to reduce stagnation in the population.
23.5.4 Takeoiler time
Tckeolvr tirize refers to the number of generations required for an evolutionary
algorithm operating under selection alone (i.e. no other operators such as
mutation or crossover) to converge to a population consisting entirely of
instances of the optimal individual, starting from a population that contains a
single instance of the optimal individual. Goldberg and Deb (1991) show that,
assuming Q> = f , the takeover time r in a population of size p for proportional
selection is
plnp- 1
t]=
TEAM LRN
c
I80
c
for .f.(x) = exp(ox). Goldberg and Deb compare these results with several other
selection mechanisms and show that the takeover time for proportional selection
(without fitness scaling) is larger than for many other selection methods.
References
Back T I994 Selective pressure in evolutionary algorithms: a characterization of selection
mechanisms Proc. 1 s t IEEE Int. Car$ on E,~oliitioiieii-yCoitipictcition (Orltindo, FL,
Jirrie 1994) (Piscataway, NJ: IEEE) pp 57-62
Baker J E 1987 Reducing bias and inefficiency in the selection algorithm Pmc. 2nd Irtt.
Cot$ o t i Genetic A1gorithrn.s (Cctmhridge, MA, 1987) ed J Grefenstette (Hillsdale,
NJ: Erlbaum) pp 14-21
de la Maza M and Tidor B 1993 An analysis of selection procedures with particular
attention paid to proportional and Boltzmann selection Proc. 5th Int. Cot$ o i i
Genetic Algorithms ( UrbaiiN-Chctmyciign,IL, July 1993) ed S Forrest (San Mateo.
CA: Morgan Kaufmann) pp 124-31
Gillies A M 1985 Mwhiiie Leciriiirig ProcwIures jhr Generciting Iiticige Dotttairi Fecitiire
Detectors Doctoral Dissertation, University of Michigan, Ann Arbor
Goldberg D E 1989 Genetic Algorithnis in Sectrch, Optimizcition, Litid Machine Leciriiiiig
(Reading, MA: Addison-Wesley
Goldberg D and Deb K 1991 A comparative analysis of selection schemes used in
genetic algorithms Foirnckitions of Genetic Algoritlrtns ed G Rawlins (San Mateo,
CA: Morgan Kaufmann) pp 69-93
Gretenstette J 1986 Optimization of control paranieters for genetic algorithms IEEE
T r m s . Syst. M m Cyhenzet. SMC-16 122-8
-1
99 1 Conditions for implicit parallelism Founrlciriotr.s ($Getietir. AIgorithm.~ed G
Rawlins (San Mateo, CA: Morgan Kaufmann) pp 252-61
Holland J H 1975 Adqmtioti iii Neitimil ctnd Art$c*icil Systems ( A m Arbor, MI:
University of Michigan Press)
Muhlenbein H and Schlierkamp-Voosen D 1993 Predictive models for the breeder genetic
algorithm E\dict. Compirt. 1 2 5 4 9
TEAM LRN
24
Tournament selection
Tobias Blickle
182
Tournament selection
(24.1)
Properties
I83
(24.2)
p ( s ) d x is the
where p ( F ) is the continuous form of p ( P ) , R ( F ) =
cumulative continuous fitness distribution and F o ( P ) < F 5 F , , , p , ( P ) the
range of the distribution function p ( F ) .
24.4 Properties
24.4. I
Conc*ateriationof toiiniamerzts
The takeover time was introduced by Goldberg and Deb (1991) to describe the
selection pressure of a selection method. The takeover time T * is the number of
generations needed under pure selection for a initial single best-fit individual to
fill up the whole population. The takeover time can. for example, be calculated
combining (24.1) and (24.3) as follows. Only the best individual is considered
and its expected proportion pie,, after tournament selection can be obtained as
= 1 /A and
pic\t = 1 - (1 - 1 /A)Y, which is a special case of (24. I ) using
Rb,-\[ = I . Performing t such tournaments subsequently with no recombination
in between leads to $best = 1 - ( 1 - I/h)qr by repeatedly applying (24.3).
Goldberg and Deb (1991) solved this equation for t and gave the following
approximation for the takeover time:
(24.4)
Figure 24.1 shows the dependence of the takeover time on the tournament size
4. For scaling purposes an artificial population size of h = e is assumed, such
that (24.4) simplifies to t:,,,(q)
ITEAM
/ Inq.LRN
Tournament selection
184
The selection intensity is another measure for the strength of selection which
is borrowed from population genetics. The selection irrturzsity S is the change
in the average fitness of the population due to selection divided by the mean
variance of the population before selection 0 , that is, S = ( U * - u)/a, with
11 average fitness before selection. and U * average fitness after selection. To
eliminate the dependence of the selection intensity on the initial distribution
one usually assumes a Gaussian-distributed initial population (Muhlenbein and
Schlierkamp-Voosen 1993). Under this assumption, the selection intensity of
tournament selection is determined by
1.5
10
15
20
25
, q
33
Figure 24.1. The selection intensity S, the loss of diversity 0 , and the takeover time r*
(for h = e) of tournament selection in dependence cm the tournament size 4.
The known exact solutions of the integral equation (24.5) are given in
table 24.1. These values can also be obtained using the results of the order
statistics theory (Back 1995). The following formula was derived by Blickle
and Thiele (199%) and approximates the selection intensity with a relative error
of less than 1 % for tournament sizes of y > 5:
TEAM LRN
185
References
Table 24.1. Known exact values for the selection intensity of tournament selection.
4
During every selection phase bad individuals are replaced by copies of better
ones. Thereby a certain amount of 'genetic material' contained in the bad
individuals is lost. The loss of diversity 8 is the proportion of the population
that is not selected for the next population (Blickle and Thiele 199%). Baker
( 1989) introduces a similar measure called 'reproduction rate, RR'. RR gives the
percentage of individuals that is selected to reproduce, hence RR = 100(1 - 0 ) .
For tournament selection this value computes to (Blickle and Thiele 199%)
References
Back T 1994 Selective pressure in evolutionary algorithms: a characterization of \election
mechanisms Proc. 1st IEEE Corif on E\dutionun Conipututiori (Orlutido, FL, Jurie
1994) (Piscataway, NJ: IEEE) pp 57-62
-I
995 Generalized convergence models for tournament- and ( p . A)-selection Proc.
6th Itit. Conj: on Genetic Algorithms (Pittshitrg, PA, Jiily 1995) ed L J Eshelman
(San Mateo, CA: Morgan Kaufmann) pp 2-8
Baker J E 1987 Reducing bias and inefficiency in the selection algorithm P m c . 2tid I t i t .
Con$ o t i Genetic Algorithrirs (Cumhridge, MA, 1987) ed J J Gref'enstette (Hillsdale,
NJ: Erlbaum) pp 14-21
-1
989 An Atiulvsis of the Effects of Selection iti Getirtic Algorithtm PhD Thesis.
Graduate School of Vanderbilt University, Nashville, TN
Blickle T and Thiele L 1995a A Coniparisori of Selection Schenic~si ~ w diri Gericvic'
Algorithnts Technical Report I I , Computer Engineering and Communication
Networks Lab (TIK), Swiss Federal Institute of Technology (ETH) Zurich
--I
99% A mathematical analysis of tournament selection Pmc. 6th lttt. Cor!f: on
Genetic Algorithms (Pittshurg, PA, July 1995) ed L J Eshclman (San Mateo, CA:
Morgan Kaufmann) pp 9-16
de la Maza M and Tidor B 1993 An analysis of selection procedure\ with particular
attention paid to proportional and Boltzmann selection P r w . 5 t h I t i t . Cot$ o t i
Gertetic Algorithms (Urhaticc-Chunipuigti, /L, Jirly 1993) ed S Forrest (San Mateo,
TEAM LRN
CA: Morgan Kaufmann) pp 124-31
186
Tournament selection
TEAM LRN
25
Rank-based selection
John Grefenstette
25.1 Introduction
Selection is the process of choosing individuals for reproduction or survival in
an evolutionary algorithm. Rank-based selectiorz or rcinking means that only
the rank ordering of the fitness of the individuals within the current population
determines the probability of selection.
As discussed in Chapter 23, the selection process may be decomposed into
distinct steps:
( i ) Map the objective function to fitness.
( i i ) Create a probability distribution based on fitness.
(iii) Draw samples from this distribution.
Ranking simplifies step (i), the mapping from the objective function .f to
the fitness function @. All that is needed is
@ ( a ; )= 6 f ( c i ; )
187
188
where (Yra[,h is the number of offspring allocated to the worst individual. The
sum of the selection probabilities is then
It follows that (;Yranh = 2 -prank, and 1 5 Prarlh 5 2. That is, the expected number
of offspring of the best individual is no more than twice that of the population
average. This shows how ranking can avoid premature convergence caused by
'super' individuals.
Prsq-ranh ( 1 =
+ [rank(i)'/(p
TEAM LRN
l)'](b - a )
( p ,A), ( p
189
(25.I )
for a suitable normalization factor c. Both of the latter methods strongly bias
the selection toward the best few individuals in the population, perhaps at the
cost of premature convergence.
25.4
( p , A), ( p
0
1/TP
if rank(i) < ( I
otherwise.
T)p
I90
Theory
The theory of rank-based selection has received less attention than the
proportional selection method, due in part to the difficulties in applying the
schema theorem to ranking. The next subsection describes the issues that arise
in the schema analysis of ranking, and shows that ranking does exhibit a form
of implicit parallelism. Characterizations of the selective pressure of ranking
are also described, including its fertility rate, selective differential, and takeover
time. Finally, a simple substitution result is mentioned.
25.5. I
The use of rank-based selection makes it difficult to relate the schema theorem
to the original objective function, since the mean observed rank of a schema is
generally unrelated to the mean observed objective value for that schema. As
a result, the relative target sampling rates (see Section 23.5.1) of two schemata
under ranking cannot be predicted based on the mean objective values of the
schemata, in contrast to proportional selection. For example, consider the
following case:
where
U I ,U - I , (1s E
~ 2 . E~ Hz.
3
HI
Assume that the goal is to maximize the objective function f . Even though
~ ( H I=) 20 > 10 = f ( H z ) , ranking will assign a higher target sampling rate
to H: than to H I .
However, ranking does exhibit a weaker form of implicit parcrllu1i.m.
meaning that it allocates search effort in a way that differentiates among a
large number of competing areas of the search space on the basis of a limited
number of explicit evaluations of knowledge structures (Grefenstette 199 1 ). The
following definitions assume that the goal is to maximize the objective funcrion.
A fitness function @ is called monotoriic- it
@([I,
* f<./1 I
5 @(q
)
f(L1,).
That is, a monotonic fitness function does not reverse the sense of any pairwise
ranking provided by the objective function. A fitness function is called strict&
morrototiic if it is monotonic and
TEAM LRN
191
Theory
A strictly monotonic fitness function preserves the relative ranking of apy two
individuals in the search space with distinct objective function values. Since
@ ( a , )= S f ( a ; ) ,ranking uses a strictly monotonic fitness function by definition.
Likewise, a selection algorithm is called momtonic if
where tsr(a) is the target sampling rate, or expected number of offspring, for
individual U . That is, a monotonic selection algorithm is one that respects
the s i r n ~ i i ~ a l - ~ , f - t h e - ~principle.
tfe.~f
A selection algorithm is called stric-tly
rnomtonic if it is monotonic and
@((a,)
< @(uJ)
individuals with better fitness values. Linear ranking selection and proportional
selection are both strictly monotonic, whereas threshold selection is monotonic
but not strict, since it may assign the same number of offspring to individuals
with different fitness values.
Finally, an evolutionary algorithm is called admissi6le if its fitness function
and selection algorithm are both monotonic. An evolutionary algorithm is sfricf
iff its fitness function and selection algorithm are both strictly monotonic.
Now, consider two arbitrary subsets of the solution space, A and B .
sorted by objective function value. By definition, B partially doniiiinfes A
( A -i B ) at time t if each representative of B is at least as good as the
corresponding representative of A. The following theorem (Grefenstette I99 I )
partially characterizes the implicit parallelism exhibited by ranking (any many
other selection methods):
Rank-based selection
192
.F=I--
B-1
4
where ap is the standard deviation of the fitness values in the population, and
I is a value called the selection intensity. Back (1995) quantifies the selection
intensity for general ( p ,A) selection as follows:
TEAM LRN
193
where Z I : Aare order statistics based in the fitness of individuals in the current
population. That is, I is the average of the expectations of the p best samples
taken from iid normally distributed random variables Z . This analysis shows
that I is approximately proportional to A / p , and experimental studies confirm
this relationship (Back 1995, Miihlenbein and Schlierkamp-Voosen 1993).
25.5.4 Takeover time
Takeo/ier tirne refers to the number of generations required for an evolutionary
algorithm operating under selection alone (i.e. no other operators such as
mutation or crossover) to converge to a population consisting entirely of
instances of the optimal individual, starting from a population that contains
a single instance of the optimal individual. According to Goldberg and Deb
(1991), the approximate takeover time t in a population of size p for rankbased selection is
In p h ( l n p )
t =
In 2
prank
= 2 and
I94
References
Bick T I995 Generalized convergence models tor tournament- and ( p ,A)-selection PIOC..
6th Int. Cot!$ o i i Genetic Algorithriis (Pittshitrgh. PA, J u l y 1995) ed L J Eshelnian
(San Mateo. CA: Morgan Kaufmann) pp 2-8
Back T and H-P Schwefel 1993 An overview of evolutionary algorithms for paramcter
optimization Ekditt. Coinpitt. 1 1-23
Baker J I985 Adaptive selection methods for genetic algorithms Pmc. 1st h i t . Car$ oii
Genetic Algorithms (Pittshitrgh, PA, July I9851 cd J J Grefenstette (Hillsdale, NJ:
Lawrence Erlbaum) pp 101-1 I
-I987 Reducing bias and inefficiency in the selection algorithm Proc. 2nd lrrt. CoiIf
o t i Getretic- A1,qorithtii.s (Cmtihridge, MA, 1087) ed J J Grefenstette (Hillsdale, NJ:
Erlbaum) pp 14-21
___ 1989 Aiiri1J~si.s
(J'tlre &fects of'Selc~choirin Genetic-Algorithms Doctoral Dissertatlon.
Department o f Computer Science, Vanderbilt University
Blickle T and Thiele L 1995 A mathematical analysis o f tournament selection Proc. 6th
Irrr. Car$ oii Getretic Algorithnis (Pittshiirgli, PA, Jitly 1995) ed L Eshclman ( San
Mateo, CA: Morgan Kaufmann) pp 9-16
Goldberg D and Deb K 1991 A cotnparative analysis of selection schemes used in
genetic algorithms Fountltitions ($Genetic Algoritlinis ed G Rawlins (San Mateo,
CA: Morgan Kaufmann) pp 69-93
Grefenstette J 199 1 Conditions for implicit parallelism FoioidLitions o$Generic- Algorit~rms
ed G Rawlins (San Mateo, CA: Morgan Kaufmann) pp 252-61
Muhlenbein H and Schlierkamp-Voosen D 1993 Predictive models for the breeder genetic
algorithm E\*olict. C'oiiipt. 1 25-49
Sc h we fe I H - P I 9 77 Nit r r r tJrisc tie 0 1 7 tirn iuriI I rg I'on Corny ii ter-ModeI f en in itte I s tle r
E~~olirtioti.s.str~rte,~ic.
( / i i t r n l i . s t ~ i ~ ~ l i tSjxteni
i t t r ~ ~ Rcr.setrrc.h 26) (Base]: Birkhauser)
____ 1987 Collective phenoinena in evolutionary systems Prepririts of' the 3 1st h i r .
Meetitig Iiitertrutiotid Societj- jt)r Gerierril Sj:~tem.sResectriA (Bicrltrpest) vol 2,
pp 1025-33
Shapiro S C (ed) I990 Enc-ycloprdirr o j Artijicid l~rtelligenc*evol 1 (New York: Wiley)
Whitley D 1989 The GENITOR algorithm and selective pressure: why rank-b.ised
allocation o f reproductive trials is best Proc. 3rd Iiit. Coizj: o i i Genetic A1gorirhri.s
(Fliirjiix, VA, Jiiire 1989) ed J Schaffer (San Matzo, CA: Morgan Kaufmann) pp I 1621
Whitley D and Kauth J 1988 GENITOR: a different genetic algorithm Pro(,. R O ~ A J .
Moitnttiin Col$ o i i Artificicil Iiitrlligrriw ( D e i i \ ~ r CO)
,
pp 1 1 8-30
TEAM LRN
26
Boltzmann selection
Sarnir W Mahfoud
26.1
Introduction
(26.1)
where T is temperature and .f; is the energy, cost. or objective function value
(assuming minimization) of solution i . Slight variations of the Boltzmann trial
exist, but all variations essentially accomplish the same thing when iterated (the
winner of a trial becomes solution i for the next trial): at fixed T , given a
sufficient number of Boltzmann trials, a Boltzmann distribution arises among
the winning solutions (over time). The intent of the Boltzmann trial is that at
high T , i and j win with nearly equal probabilities, making the system fluctuate
wildly from solution to solution; at low T , the better of the two solutions nearly
always wins, resulting in a relatively stable system.
TEAM LRN
I95
196
Boltzmann selection
'I-'!
' 8 '
198
Boltzmann selection
size
P (0) c initialize-population()
T ( 1 ) c initialize-temperature()
for t +- 1 to g do
P ( t ) t shuffle(P(t - I ) )
for i t 0 to p / 2 - I do
PI +- W J + l ( t )
P2 +- a2,+2(t)
{ c l , c . 2 ) t recombine(p1, p z )
t neighborhood(c.1)
ci t neighborhood(c.2)
if random() > [ I + e ( l ) - ~ ( ~ ) l ~ then
( f ) l -alz , + ~ ( t t
) c, fi
if random() > [ I + el f(/>)-!(;)I/r(f)
] - I then U ? ~ + Z (+-~ )c$ fi
od
T ( t 1 ) t adjust-temperature()
od
(9;
199
i
j
A:B
C:D
A : D
C : B
1
Figure 26.1. The population, after application of crossover and mutation (step 1 ),
transitions from superstring i to superstring j . After a Boltzmann trial (step 2), either
i or j becomes the current population. Individual population elements are represented
as rectangles within the superstrings. Blocks A, B, C, and D represent portions of
individual population elements, prior to crossover and mutation. Crossover points are
shown as dashed lines. Blocks A, B, C, and D result from applying mutation to A, B.
C, and D.
200
Boltzmann selectic2
operator meets certain conditions. According to Aarts and Korst (1989), two
conditions on the neighborhood generation mechanism are sufficient to guarantee
asymptotic global convergence. The first condition is that the neighborhood
operator must be able to move from any state to a globally optimal state in a finite
number of transitions. The presence of mutation satisfies this requirement. The
second condition is symmetry. It requires that the probability at any temperature
of generating state y from state .r is the same as the probability of generating
state .r from state 3. Symmetry holds for common crossover operators such as
single-point, multipoint, and uniform crossover (Mahfoud and Goldberg 199s ).
References
Aarts E and Korst J 1989 Siinitkitetl Atinediiig titid Holtmciiin Mtrchines: CI Stochastic
Appro(rch to Comhirrtrtoricrl 0prinii;crtinn crnd Nritrrrl Corripiiriricq (Chichest3r:
Wiley)
Azencott R (ed) I992 Simitkitt~dAniir-.diiig: Ptiriillelizcrtioii Techiiiyites (New York:
Wiley)
de la Maza M and Tidor B 1993 An analysis of selection procedures with particular
attention paid to proportional and Boltzmann selection Proc. 5th Int. Car$ o t i
Getretic Algorithms ( Ur~~urici-C~i~tnip~tigti,
lL, Jrtly IY93) ed S Forrest (San Matco.
CA: Morgan Kaufmann) pp 124-31
Goldberg D E 1990 A note on Boltzmann tournament selection for genetic algorithms
and po pu 1at i o n - or ie n t ed si m u 1ated anneal i ng Conzplex Sjw. 4 445-60
Ingber L and Rosen B I992 Genetic algorithms and very fast simulated re-annealing: a
comparison Mcith. Cotitpt. Modelling 16 87- 100
Kirpatrick S, Gelatt C D Jr and Vecchi M P 1983 Optimization by simulated annealing
Science 220 67 1-80
Mahtoud S W 1993 Finite Markov chain models of an alternative selection strategy tor
the genetic algorithm Conip1e.v Syst. 7 155-70
-1995 Nithirig Methods fiw Crtietic Algorithms Doctoral Dissertation and IlliG.4L
Report 9500 I . UniLwsity of Illinois at Urhana-Champaign. Illinois Gencstic
Algorithms Laboratory; Dissertcition Ahstrctcts Int. 56(9) p 49878 (Uni\vr:virx
Mii.ro~1rii.vY543663)
Mahtoud S W and Goldberg D E 1992 A genetic algorithm for parallel simulated
annealing Proc. 2nd lilt. Col$ oti Partillel Prohleni Sohirig from Nmire (Briissols,
1992) ed R Manner and B Manderick (Amsterdam: Elsevier) pp 301-10
-1995 Parallel recombinative simulated annealing: a genetic algorithm Ptrrtillel
cor?lpllt. 2 1 1 -2 8
Romeo F and Sangiovanni-Vincentelli A 199I A theoretical framework for simulated
annealing Algorithniicvi 6 3 0 2 4 5
TEAM LRN
27
Other selection methods
David B Fogel
27.1
Introduction
20 I
202
where .f(z)is the objective value of the solution z and f ( t ) is the mean of all
solutions in the population at time t . Thus a solutions fitness increases with
its distance from the mean of all current solutions. The idea is to distribute
more search effort to both the extremely good and extremely bad solutions. The
utility of this method is certainly very problem dependent.
27.5
Boltzmann selection
203
Competitive selection
+ 1 /popsize
27.7
Competitive selection
27.8
Variable lifespan
Finally, Back (1996) notes that the concept of a variable lifespan has been
incorporated into the ( p . A ) selection of evolution strategies by Schwefel and
Rudolph ( 1995) by allowing the parents to survive some number of generations.
When this number is one generation, the method is the familiar comma strategy:
TEAM
LRN
at infinity, the method becomes a plus
strategy.
2 0.1
References
Altcnberg L 1993, Emergent phenomena in genetic programming, Proca. -3rd Anti. Cot:/:
o t i E\-olirtiotiiit;v Pro,qrtiriitiiiticq ( S m Diego. CA, Fohriuit;v 1994) cd A V Sebald arid
1- J Fogcl (Singapore: World Scientific) pp 233-1 1
__I
OC)3b The c\olution o f evolvability in genetic programming Atl\*ciricx~si t i Getic~ric.
~ro~qrtittittiiti~q
ed K Kinnear (Cambridge, MA: MIT Press) pp 37-73
Axelrod R I987 The c\~olutiono t strategies in the iterated prisoners dilemma Grtietic.
Algoritlirtis titit1 SirtiirItitcd A t i t i t w l i t i g ed L Davis ( Los AI tos, CA: Morg;in
Kaufmann) pp 3 2 1 I
Angclinc P J and Pollack J B 1993 Competitive environments e\wlve better solutions for
complex tasks Pro(.. 5th I t i t . Cot!f: o t i Gciirtic Algorithim ( U r t ~ ~ i t i ~ i - C l i ~ i t t i/I.,
i,~ii~~~i
Jirl\s I Y Y 3 ) ed S Forrest (San Matco, CA: Morgan Kaufmann) pp 264-70
Back -I- I c)W Selective pressure in evolutionary algorithms: a characterization of se1ectic)n
riicchanisms Prot.. t.vt /EEE Corlf: O I I E \ d l t r i o t l t i r - j , Cortipiittitiori (Orlutido, F I,,
100.3) (Piscataway, NJ: IEEE) pp 57-62
___ I 096 Et.olittioticir;\. tllgot-ithis it! Tlitwr;~~
tint1 Prwticv ( New, York:
Oxford
Llii i vers i t j , Press )
de la M w a hl and Tidor B 1993 An analysis ot selection procedures with particular
attention paid to proportional and Boltmiann selection Proc.. 5th I t i t . Cot!/: w i
G c v i t h c AI~qoritlit~rs
I ~ l r ~ , ~ i t i t i - C l i ( i t i i i , IL,
( i i ~Jirly
~ ~ i ~19Y-3) ed S Forrest (San Mattw,
CA: Morgun Kauft-nann) pp 124-3 1
FogeI D B I995 E\~olirtioritir-y Coriil,irtcctioti: To\rwtd ( I N r ~ yPhilosophy of Mtic~hiric~
I r r t c ~ l l i ~ q e t i c(~Nc ~~ NYork:
,
IEEE)
Fogel L J and Burgin G H 1969 C o r i i p c t i t i r v God-Seeking TIiroiqIi E\wliitioticrr:\.
~ t . o ~ q t . ~ i t i i t t i iFinal
t i ~ ~ Report, Contract N o AF I9(628 )-5927, Air Force Cambridge
Re sc arch Lab r at or i e s .
Hillis b D 1992 Co-c\~ol\~ing
parasites impro\,es simulated evolution as an optimintion
procedure Arrificicil Lit;. 11 ed C Langton, C Taylor, J Farmer and S Rasmus?.cn
(Reading, MA : Addison-Wesley ) pp 3 13-24
Kuo T and H ~ m S-Y
g 1993 A genetic algorithm with disrupti\re selection Pro(.. 5th / t i t .
Cot!f: o t i Gctietic. Algot-itlitiiv ( L ~ rh t i t i t i - C h t i t i i i ~ c i i ~ qt
It,.i , J i i / y 1993) ed S Forrest (San
Matco, CA: Morgan Kaufmann) pp 65-9
Michalewici Z I996 Grtiotic~Al,qorit/ittis + Dtitti Strrrctirrus = E\diitiori Pro,qrrini.s . h d
cdn (Berlin: Springer)
Sch\+vtcI H-P and Riidolph G I995 Contemporary evolution strategic5 , 4 t / i ~ t r i i t ~in
.s
A rt(fic.iti1 l.(fi)( Pt.oc.. 3rd Itit. Cot!$ o t i t\t,t(f;citil Lifr, G r m t i h , S p r i t i ) ( Loctiirr N c t c s
i t t Artific.iril ltitolligctic~cY 2 Y ) ed 1: Morrin r t trl (Berlin: Springer) pp 893-907
Scbald A V and Schlen/ig J 1094 Minirnax design of neural nel conlrollcrx for highly
.~
Nc>t\t.ork.\ NN-5 73-82
uncertain plant5 IE>EE7 k ~ t iNeirrril
TEAM LRN
28
Generation gap methods
Jayshree Sarrna and Kenneth De Jorzg
28.1 Introduction
The concept of a generation gap is linked to the notion of riorio,.er-ltrpl,iriS and
o\verlappiizg populations. In a nonoverlapping model parents and offspring nei er
compete with one another, i.e. the entire parent population is always replaced by
the offspring population, while in an overlapping system parents and offspring
compete for survival. The term generation gap refers to the amount of o\erlap
between parents and offspring. The notion of a generation gap i \ closely related
to selection algorithms and population management issues.
A selection algorithm in an evolutionary algorithm (EA) involves two
elements: (i) a selection pool and (ii) a selection distribution over that pool.
A selection pool is required for reproduction selection as well as for deletion
selection. The key issue in both these cases is 'what does the pool contain urhen
parents are selected and when survivors are selected'?'.
In the selection for the reproduction phase, parents are selected to produce
offspring and the selection pool consists of the current population. Hou the
parents are selected for reproduction depends on the individual EA paradigm.
In the selection for the deletion phase, a decision has to be made as to
which individuals to select for deletion to make room for the new offspring.
In nonoverlapping systems the entire selection pool consisting of the current
population is selected for deletion: the parent population (,U) is always replaced
by the offspring population (A). In overlapping models, the selection pool for
deletion consists of both parents and their offspring. Selection for deletion is
performed on this combined set and the actual selection procedure varies in each
of the EA paradigms.
Historically, both evolutionary programming and evolution strategies
had overlapping populations while the canonical genetic algorithms used
nonoverlapping populations.
TEAM LRN
206
207
(identified empirically for the test suite used) meant that approximately 40% of
the offspring were clones of their parents, even for G = 1 . A later empirical
study by Grefenstette ( 1986) confirmed the earlier results that a larger generation
gap value improved performance.
However. early experience with classifier systems (e.g. Holland and Reitman
1978) yielded quite the opposite behavior. In classifier systeiiir only a w h e t of
the population is replaced each time step. Replacing a small number of classifiers
was generally more beneficial than replacing a large number or possibly all of
them. Here the poor performance observed as the generation gap L alue increased
was attributed to the fact that the population ar a whole represented a Gngle
solution and thus could not tolerate large changes i n it\ content.
In recent years, computing equipment with increased capacity is easily
available and this effectively removes the reason for preferring the RJ approach.
The desire to solve more complex problems using genetic algorithms has
prompted researchers to develop an alternative to the generational sy stem called
the 'steady state' approach, in which typically parents and offspring do coexist
(see e.g. Syswerda 1989, Whitley and Kauth 1988).
208
Figure 28.2. The mean and variance of the growth curves of the best in a nonoverlapping
system (population size, 50; G = 1).
209
only about 80% of the time and the growth curve, exhibit much higher variance
when compared to the nonoverlapping population (figure 28.2).
This high variance for small generation gap values causes more genetic
drift (allele loss). Hence, with smaller population sizes, the higher variance in
a steady state system makes it easier for alleles to disappear. Increasing the
population size is one way to reduce the the variance (see figure 28.3) and thus
offset the allele loss. In summary, the main difference between the generational
and steady state systems is higher genetic drift in the latter especially when
small population sizes are used with low generation gap values. (See the article
by De Jong and Sarma (1993) for more details.)
Figure 28.3. The mean and variance of the grouth curve5 of the best in an overlapping
5ystem (population w e , 200; G = 1/200)
2 10
References
21 1
References
Back T, Hoffmeister F and Schwefel H-P 1991 A survey of eLrolutionary strategies Pro(..
4th liit. Coi$ o i i Gerietic Algorithms (Sun Diego, CA, J u I j . 1991I ed R K Belew and
L B Booker (San Mateo, CA: Morgan Kaufmann) pp 2-9
De Jong K A 1975 Ati Aiitr1y.si.sof the Behavior cf ei Clris.s of' Gerirtic Acltipti\>t>Sjsteiiis
Phd Dissertation, University of Michigan
-1
993 Genetic algorithms are NOT function optiniizcrs Foiiritltitioiis of' Geiietic
Algorithms 2 ed L D Whitley (San Mateo, CA: Morgan Kaufmann) pp 5-17
De Jong K A and Sarma J 1993 Generation gaps revisited Foiiiihtions of' Genrtic
Algorithms 2 ed L D Whitley (San Mateo, CA: Morgan Kaufmann) pp 19-28
Fogel G B and Fogel D B 1995 Continuous evolutionary programming: analysis and
experiments Cylwrii~t.Syst. 26 79-90
Fogel L J, Owens A J and Walsh M J 1996 Art(ficiti1 Iiitelli,qcwc~c~
through Siiiiuleitcd
Eidirtioii (New York: Wilcy)
Goldberg D E and Deb K 1991 A comparative analysis of selection schemes used in
genetic algorithms Foiiiidcitiniis of Genetic Alp~r-ithnisI ed G J E Rawlins (San
Mateo, CA: Morgan Kaufmann) pp 69-93
Grefenstette J J 1986 Optimization of control parameters for genetic algorithms / E
Truns. Syst. Mtiii Cj)heriiet. SMC-16 122-8
Holland J H 1975 Adtiptcitioti in Nuturul ciiid Art$citrl Systenis ( Ann Arbor, MI:
University of Michigan Press)
Holland J H and Reitman J S 1978 Cognitive systems based on adaptikte algorithms
Pcrtterti-Directed li!forenc.e Sj,steiii.s ed D A Waterman and F Hayes-Roth (New
York: Academic)
Peck C C and Dhawan A P 1995 Genetic algorithms as global random search methods:
an alternative perspective E~dutioiiriq~
Contpiit. 3 39-80
Rechenberg I 1 973 Ei~olutioiis.strtrtegir: 0priiiiic.r-irng tec*/irii.vc~hc~r
S>atoriir i i r i d i
Priiizipieii der hiologischeti \dutioti (Stuttgart: Frommann-Holzboog )
Schwefel H-P 198 1 Nuinerical 0ptinii:atioii qf Coinputer Moclels (Chichester: Wiley )
Syswerda G 1989 Uniform crossover in genetic algorithms P roc. 3rd I t i t . CorJf: o r i Gerwtic.
Algorithms (Fciiifkv, VA, Juiie 1989) ed J D Schaffer (San Mateo, CA: Morgan
Kaufmann) pp 2-9
-199
1 A study of reproduction in generational and steady-state genetic algorithms
Foirnc1citioii.s cf Gerietic Algorithnzs I ed G J E Rawlins (San Mateo. CA: Morgan
Kaufmann) pp 94-101
Whitley D 1989 The GENITOR algorithm and selection pressure: why rank-based
allocation of reproductive trials is best Proc. 3rd liit. Cmf: oil G o i i d ( .AI,qorithi?i.s
(Ftiirjkv, VA, J i i i i e 1989) ed J D Schaffer (San Mateo, CA: Morgan Kaufmann)
pp 116-21
Whitley D and Kauth J 1988 GENITOR: a Di'ereitt
Gcwtic Algorithni Colorado State
University Technical Report CS-88- 10 1
TEAM LRN
29
A comparison of selection mechanisms
Peter J B Hancock
29.1
Introduction
212
TEAM LRN
Simulat ions
213
This captures the notion that it is harder to produce a given step in average
fitness between the population and those selected when the fitness variance is
low. However, both takeover time and selection intensity depend on the fitness
functions, and so theoretical results may not always transfer to a real problem.
There is an additional difficulty because the fitness variance itself depends on
the selection method, so different methods configured to have the same selection
intensity may actually grow at different rates.
Most of the selection schemes have a parameter that controls either the
proportion of the population that reproduces or the distribution of reproductive
opportunities, or both. One aim in what follows will be to identify some
equivalent parameter settings for different selection methods.
29.2
Simulations
2 14
the population. Some selection methods are better at preserving such diversity:
other things being equal, this seems likely to improve the quality of the overall
search (Muhlenbein and Schlierkamp-Voosen 1095, Blickle and Thiele 199%).
It should be emphasized that fast convergence on these tasks is not
necessarily good: they are deliberately simple in an effort to illustrate some
of the differences between selection methods and the reasons underlying them.
Good selection methods need to balance exploration and exploitation. Before
reporting results, we shall consider a number of more theoretical points of
similarities and differences.
215
Goldberg and Deb (1991) showed that simple binary tournament selection (TS)
(see Chapter 24) is equivalent to linear ranking (Section 25.2) when set to give
two offspring to the top-ranked string (prank = 2). However, this is only in
expectation: when implemented the obvious way, picking each fresh pair of
potential parents from the population with replacement, tournament selection
suffers from sampling errors like those produced by roulette wheel sampling,
precisely because each tournament is performed separately. A way to reduce
this noise is to take a copy of the population and choose pairs for tournament
from it without replacement. When the copy population is exhausted, another
copy is made to select the second half of the new population (Goldberg ct crl
1989). This method ensures that each individual participates in exactly two
tournaments, and will not fight itself. It does not eliminate the problem, since.
for example, an average individual, that ought to win once, may pick better or
worse opponents both times, but it will at least stop several copies of any one
being chosen.
The selection pressure generated by tournament selection may be decreased
by making the tournaments stochastic. The equivalence, apart from sampling
errors, with linear ranking remains. Thus TS with a probability of the better
string winning of 0.75 is equivalent to linear ranking with rtjrank = 1.5. The
selection pressure may be increased by holding tournaments among more than
two individuals. For three, the best will expect three offspring, while an
average member can expect 0.75 (it should win one quarter of its expected
three tournaments). The assignment is therefore nonlinear and Back ( 1994)
shows that, to a first approximation, the results are equivalent to exponential
nonlinear ranking, where the probability of selection of each rank i, starting at
i = 1 for the best, is given by (s - l)(s'-')/(d' - l ) , where s is typically in the
range 0.9-1 (Blickle and Thiele 199%). (Note that the probabilities as specified
by Michalewicz (1992) do not sum to unity (Back 1994).) More precisely,
they differ in that TS gives the worst members of the population no chance to
reproduce. Figure 29.1 compares the expected number of offspring for each rank
in a population of 100. The difference results in a somewhat lower population
diversity for TS when run at the same growth rate.
Goldberg and Deb (1991) prefer TS to linear ranking on account of its
lower time complexity (since ranking requires a sort of the population), and
Back (1994) argues similarly for TS over nonlinear ranking. However, time
complexity is unlikely to be an issue in serious applications, where the evaluation
time usually dominates all other parts of the algorithm. The difference is in any
case reduced if the noise-reduced version of TS is implemented, since this
also requires shuffling the population. For global population models, therefore,
ranking, with Baker's sampling procedure (Baker I987), is usually preferable.
TS may be appropriate in incremental models, where only one individual is
to be evaluated at a time, and in TEAM
parallelLRN
population models. It may also be
216
25
50
75
I00
Rank of individual
Figure 29.1. Expected number o ! otfspring against rank for tournament selection
tournament s i x 3 and exponential rank selection with J = 0.972.
U.
ith
appropriate in, for instance, game playing applications, where the evaluation
itself consists of individuals playing each other.
Freisleben and Hiirtfelder (1993) compared a number of selection schemes
using a meta-level GA. that adjusted the parameters of the GA used to tackle
their problem. Tournament selection was chosen in preference to rank selection,
which at first sight seems odd, since the only difference is added noise. A
possible explanation lies in the nature of their task, which was learning the
weights for a neural net simulation. This is plagued with symmetry problems
(e.g. Hancock 1992). The GA has to break the symmetries and decide on
just one to make progress. I t seems possible that the inaccuracies inherent in
tournament selection faci I i tated this symmetry breaking, with one individual
having an undue advantage, and thereby taking over the population. Noise is
not always undesirable. though there may be more controlled ways to achit:ve
the same result.
217
error. Incremental models also suffer in the presence of evaluation noise (see
Section 29.6).
Generational, hest
Generational, worst
Incremental, hest
Incremental, worst
Evaluations
Figure 29.2. The growth rate in the presence of mutation of the best and worst in the
population for the incremental model with random deletion and the generational model.
both with linear rank selection for reproduction, PrClnh
= 1.2.
Some GA workers allow only the top few members of the population to
reproduce (Nolfi et nl 1990, Muhlenbein and Schlierkamp-Voosen 1993). This
is often called truncation selection, and is equivalent to the ES ( p , A) approach
subject only to a difference in what is called the population (see Section 29.3).
EP uses a form of tournament selection where all members of the extended
population y h compete with c. others, chosen at random with replacement.
Those y that amass the most wins then reproduce by mutation to form the next
extended population. This may be seen as a rather softer form of truncation
selection, converging to the same result as a ( p y ) ES as the size of c
increases. The value of c does not directly affect the selection pressure, only
the noise in the selection process.
The EP selection process may be softened further by making the tournaments
probabilistic. One approach is to make the probability of the better individual
winning dependent on the relative TEAM
fitnessLRN
of the pair: p , = j ; / ( . f ; f , ) (Fogel
218
1988). Although intuitively appealing, this has the effect of reducing selection
pressure as the population converges and can produce growth curves remarkahly
similar to unscaled fitness proportional selection (FPS; Hancock 1994).
Simple FPS suffers from sensitivity to the distribution of fitness values in the
population, as discussed in Chapter 23. The reduction of selection pressure as
the population converges may be countered by moving baseline techniques, such
as windowing and sigma scaling. These are still vulnerable to undesirable loss
of diversity caused by a particularly fit individual, which may produce many
offspring. Rescaling techniques are able to limit the number of offspring given
to the best, but may still be affected by the overall spread of fitness values, .And
particularly by the presence of very poor individuals.
Figure 29.3 compares takeover and growth rates of FPS and some of the
baseline adjustment and rescaling methods. The simple takeover rates for the
three adjusted methods are rather similar for these scale parameters, with linear
scaling just fastest. Simple FPS is so slow it does not really show on the
same graph: i t reaches only 80% convergence after 40000 evaluations on this
problem. The curves for growth in the presence of mutation are all rather
alike: the presence of the mutation maintains the range of fitness values in the
population, giving simple FPS something to work on. Note, however, that i t
still starts off relatively fast and slows down towards the end: probably the
opposite of what is desirable. The three scaled versions are still similar, but
note that the order has reversed. Windowing and sigma scaling now grow more
rapidly precisely because they Fail to limit the number of offspring to especially
good individuals. A fortuitous mutation is thus better exploited than in the
more controlled linear scaling, which leads to the correct result in this simple
hill-climbing task, but may not in a more complex real problem.
29.5.2 Rcrrrkirig
Goldberg and Deb (1991) show that the expected growth rate for linear ranking
is proportional to the value of prank, the number of offspring given to the best
individual. For exponential scaling, the selection pressure is proportiond to
1 - s . This makes available a wide range of selection pressures, defined by the
value of s, illustrated in figure 29.4. The highest takeover rate available with
linear ranking (prank = 2 ) is also shown. Exponential ranking can go faster with
smaller values of s (see table 29. I ). Note the logarithmic x-axis on this plot.
With exponential ranking, because of the exponential assignment curve, poor
individuals do rather better than with linear ranking, at the expense of those more
TEAM
in the middle o f the range. One result
of LRN
this is that, for parameter settings that
Simulation results
219
give similar takeover times, exponential ranking loses the worse values in the
population more slowly, which may help preserve diversity in practice.
29.5.3 Evolution strategies
I00
c
.-
2n
75
I7
50
%
L
25
0
0
1000
2000
3000
4000
5000
2000
Evaluations
4000
6 000
Evaluations
Figure 29.3. (a) The takeover rate for FPS, with windowing, sigma, and inear scaling.
(b) Growth rates in the presence of mutation.
U Exp s=0.999
'
Exp s=0.998
Exps=0.996
'
A
-
Exp s=0.992
=--
Exp s=0.984
__O__
Exps=0.968
+ Linear 2.0
100
1000
10000
Evaluations
Figure 29.4. The takeover rate for exponential rank selection for a number of values of
s, together with that for linear ranking, TEAM
prank=LRN
2.
220
table 29.1). One simulation result is shown, in figure 29.5, to make clear
the selection pressure achievable by ( p , A) selection, and indicate its poteniial
susceptibility to evaluation noise, discussed further below.
Noise free
I,l00
+ 10. I00
25,100
With evaluation n o i x
__)__
I.100
+ 10,100
& 25.100
I
5000
10000
Evaluations
1
1 5000
20000
Figure 29.5. The growth rate in the presence of mutation for ES ( p ,A) selection \vith
and without evaluation noise, for h = 100 and p = 1, 10. and 25.
22 I
Simulation results
1000
2000
3000
4000
5000
6000
Evaluations
Figure 29.6. The takeover rates for the generational model and the kill-oldest incremental
model, both using linear ranking for selection.
kr 1 . 1
*
*
k r 1.2
k r 1.4
+ rI 1.2
rl 1.4
+ rI 1 . 8
" I
2500
I
5000
7500
I
10000
Evaluations
Figure 29.7. Growth rates in the presence of mutation for incremental kill-by-inverserank (kr) and generational linear ranking (rl) for various values of Drank.
222
growth rate changes more rapidly than prank. This is because an increase in p u n k
has two effects: increasing the probability of picking one of the better members
of the population at each step, and increasing the number of steps for which
they are likely to remain in the population, by decreasing their probability of
deletion. Figure 29.7 compares growth rates in the presence of mutation for
kill-by-rank incremental and equivalent generational models. It may be seen
that the generational model with prank = 1.4 and the incremental model with
prank = 1.2 produce very similar results. Another matched pair at lower gronth
rates is generational with prank = 1.2 and incremental with prank = 1.13 (not
shown).
One of the arguments in favor of incremental models is that they allow
good new individuals to be exploited at once, rather than having to wait a
generation. It might be thought that any such gain would be rather slight, since
although a good new member could be picked at once, it is more likely to
have to wait several iterations at normal selection pressures. There is also the
inevitable sampling noise to be overcome. De Jong and Sarma (1993) claim
that there is actually no net benefit, since adding new fit members has the
effect of increasing the average fitness, thus reducing the likelihood of them
being selected. However, this argument applies only to takeover problems:
when reproduction operators are included the incremental approach can generate
higher growth rates. Figure 29.8 compares the growth of an incremental killoldest model with a generational model using the same selection scheme. The
graph also shows one of the main drawbacks of the incremental models: their
sensitivity to evaluation noise, to be discussed in the following section.
223
Noise free
Kill worst
0.75
C
.-
Kill oldest
+ Generational
a
3
0.5
C
.I
Kill worst
0
v)
0.25
+ Generational
0
5 000
10000
I5000
Evaluations
Figure 29.8. Growth in the presence of mutation, with and without evaluation noise, for
the generational model with linear ranking and incremental models with kill-worst and
kill-oldest, all using prank= 1.2 for selection.
No noise
Rank
+ Sigma
With evaluation noise
+ Rank
d
2000
4000
6000
Sigma
8000
Evaluations
Figure 29.9. Growth in the presence of mutation, with and without evaluation noise, for
the generational model with linear ranking,
TEAMprank
LRN= 1.8, and sigma-scaled FPS, s = 4.
224
225
Conclusions
Table 29.1. Parameter settings that give equivalent selection intensities for ES ( E L ,A).
TS, and linear and exponential ranking, adapted and extended from Blickle and
Thiele (199%). Under tournament size, p refers to the probability of the better string
winning.
I
ES ,u/A
0.1 I
0.34
0.56
0.84
1.03
1.16
1.35
I .54
1.87
0.94
0.80
0.66
0.47
0.36
0.30
0.22
0.15
0.08
Tournament size
2, p = 0.6
2, p = 0.8
2
3
4
5
7
10
20
prank,
Lin rank
1.2
1.6
2.0
.s,
29.8 Conclusions
The choice of a selection mechanism cannot be made independently of
other aspects of the evolutionary algorithm. For instance, Eshelman ( 1991 )
deliberately combines a conservative selection mechanism with an explorative
recombination operator in his CHC algorithm. Where search is largely driven by
mutation, it may be possible to use much higher selection pressures. typical of
the ES approach. If the evaluation function is noisy, then most incremental
models and others that may retain parents are likely to suffer. Certainly.
selection pressures need to be lower in the presence of noise, and, of the
incremental models, kill-oldest fares best. Without noise, incremental methods
can provide a useful increase in exploitation of good new individuals. Care
is needed in the choice of method of deletion: killing the worst provides high
growth rates with little means of control. Killing by inverse rank or killing
the oldest offers more control. Amongst generational models, the ES ( p ,A)
and exponential rank selection methods give the biggest and most controllable
range of selection pressures, with the ES method probably most suited to
mutation-driven, high-growth-rate systems, and ranking better for slower, more
explorative searches, where maintenance of diversity is important.
References
Back T I994 Selective pressure in evolutionary algorithms: a characterization of selection
methods Proc. 1st IEEE Cot$ on Evolutionan Compiitcition (Orlnndo. FL, Jiinr
1994) (Piscataway, NJ: IEEE) pp 57-62
-1
995 Generalized convergence models for tournament and ( p . A)-selection Proc. 6th
Int. Conf on Genetic Algorithms (Pittsburgh, PA, July 1995) ed L J Eshelman (San
TEAM
Mateo, CA: Morgan Kaufmann) pp
2-8 LRN
226
References
227
TEAM LRN
30
Interactive evolution
Wougang Banzhaf
30.1 Introduction
The basic idea of interactive evolution (IE) is to involve a human user on-line
in the variation-selection loop of the evolutionary algorithm (EA). This is to be
seen in contrast to the conventional participation of the user prior to running
the EA by defining a suitable representation of' the problem (Chapters 14-21),
the fitness criterion for evaluation of individual solutions, and corresponding
operators (Chapters 31-34) to improve fitness quality. In the latter case, the
user's role is restricted to passive observation during the EA run.
The minimum requirement for IE is the definition of a problem
representation, together with a determination of population parameters only.
Search operators of arbitrary kind as well as selection according to arbitrary
criteria might be applied to the representation by the user. The process is much
more comparable to the creation of a piece of art, for example, a painting, 1han
to the automatic evolution of an optimized problem solution. In IE, the user
assumes an active role in the search process. At the minimum level, tht: IE
system must hold present solutions together with variants presently generJted
or considered.
Usually. however, automatic means of variation (i.e. evolutionary sedrch
operators using random events) are provided with an IE system. In the present
context we shall require the existence of automatic means of variation by
operators for mutation (Chapter 32) and recombination (Chapter 33) of solutions
which are to be defined prior to running the EA.
30.2 History
Dawkins (1986) was the first to consider an elaborate IE system. The evolution
of biomorphs, as he called them, by IE in a system that he had originally intended
to be useful for the design of treelike graphical forms has served as a prototype
for many systems developed subsequently. Starting with the contributions of
Sims (1991) and the book of Todd and Lathani (1992), computer art developed
into the present major application area of IE.
228
TEAM LRN
The problem
229
230
Interactive evolution
I
2
3
4
5
6
7
8
9
10
er,0,
individual last selected during the run, or
P * , the population last selected during the run.
p , A, 0,,o,
Input:
Output:
U * ,the
to;
P ( t ) t initialize(p);
while ( i ( P ( t ) ,0 , )# true) do
Input: O:, O&
P(t) t recombine( P ( t ) , Or, 0;);
P ( t ) t mutate(P(t), (HI, O&>;
output: P ( t )
Input:
P ( t 1) +- select(P(t),p , @I,,0 : ) ;
t+t+l;
od
Input:
Output:
1
2
3
4
5
6
7
8
9
10
I1
12
13
p , A, q. +I>1.
em,
o,, o r ,
O S
U * ,the
to;
P ( r ) c initialize(p);
while ( l ( P ( t ) .0,)
# true) do
Input: O:, @I;
P(r) t recombine(P(t), Or,(HI:);
P ( t ) t mutate(P(t), O m , 0;);
F ( t ) t evaluate(P(t),A);
P(r) t sort(P(t),0,);
P(t) t select(P(t),F ( t ) ,p , q , 0,);
output: P(t)
Input: (3;
P ( t I ) t select(P(t),p , OS,0;);
t c t f l ;
od
The newly added parameter @Iois used here to specify the predefined order
of the result after evaluation according to the predefined criterion. As before,
the GII):-parametersare used to specify the user interaction with the system. q
is the parameter stating how many of the automatically generated and ordered
variants are to be presented to the TEAM
user. LRN
If p A = q in a ( p A)-strategy, or
23 1
Difficulties
30.5 Difficulties
The second, more complicated version of IE requires a predefined fitness
criterion, in addition to user action. This trades one advantage of IE systems for
another: the absence of any requirement to quantify fitness for a small number
of variants to be evaluated interactively by the user.
Interactive systems have one serious difficulty, especially in connection
with the automatic means of variation that are usually provided: whereas the
generation of variants does not necessarily require human intervention, selection
of variants does call the attention of the user. Due to psychological constraints,
however, humans can normally select only from a small set of choices. IE
systems are thus constrained to present only of the order of ten choices at each
point in time from which to choose. Also in sequence, only a limited number
of generations can be practically inspected by a user before the user becomes
tired.
It is emphasized that this limitation must not mean that the generation of
variants has to be restricted to small numbers. Rather the variants have to be
properly ordered at least, for a presentation of a subset that can be handled
interactively.
232
Interactive evolutic
Table 30.1. An overview of different IE systems.
Application
Genotypic elements
Phenotype
Source
Lifelike structures
Textures. images
biomorphs
(s.y , :) pixel
values
( A . y , :) pixel
values
face images
Dawkins (1986)
Sims (1991)
3D rendering of
grown objects
Caldwell and
Johnston ( I90 1 )
Todd and
Latham ( 1992)
system behavior
Sims (1992)
rendered objects
McCormac k
(1994)
Nguyen and
Huang ( 1994)
Graf and
BanLhaf ( I 99Sa)
Animation
Person tracking
Images. sculptures
Dynamical systems
I mages, animation
Airplane design
Images, design
airplane drawings
bilmap images
Sims (1991)
References
233
Figure 30.1. Samples of evolved objects: ( U ) dynamical system, cell structure (Sims
1992, @ MIT Press); ( 6 ) artwork by Mutator (Todd and Latham 1992, with permission
of the authors); (c) hybrid car model (Graf and Banzhaf 1995b, @ IEEE Press).
References
Caldwell C and Johnston V S 1991 Tracking a criminal suspect through face-space with
a genetic algorithm Proc. 4th Int. COT$on Genetic Algorithms (San Diego, CA, July
199I) ed R K Belew and L B Booker (San Mateo, CA: Morgan Kaufmann) pp 416TEAM LRN
21
234
Interactive evolution
Further reading
This section is intended to give an overview of presently available work in IE
and modeling methods which might be interesting to use.
1. PrusinkiewicL P and Lindenmayer A I99 1 The Algorithmic Beuutj, ofPlunts (Berlin:
Springer)
An informative introduction to L-systems and their use in computer graphics.
2. Koza J R 1992 Genetic Programming (Cambridge, MA: MIT Press)
A book describing methods to evolve computer code, mainly in the form of LISPtype S-expressions.
3. Caldwell C and Johnston V 1991 Tracking a criminal suspect through 'face-space'
with a genetic algorithm Proc. Int. Con$ on Genetic Algorithms (Sun Diego, CA,
July 1991) ed R K Belew and L B Booker (San Mateo, CA: Morgan Kaufmann)
pp 416-21
Very interesting work containing one of the more profane applications of IE.
4. Baker E 1993 Evolving line drawings Proc. Int. Cot$ on Genetic Algorithms
(Urbanu-Champaign, IL, July I993) ed S Forrest (San Mateo, CA: Morgan
Kaufmann) p 627
This contribution discusses new ideas on design using simple style elements for IE.
31
Introduction to search operators
Zbignie w Michalew icz
Any evolutionary system processes a population of individuals, P ( t ) =
{ a : ,. . . , U ; ] } ( t is the iteration number), where each individual represents a
potential solution to the problem at hand. As discussed in Chapters 14-2 I , many
possible representations can be used for coding individuals; these representations
may vary from binary strings to complex data structures I .
Each solution a: is evaluated to give some measure of its fitness. Then a
new population (iteration t 1) is formed by selecting the more-fit individuals
(the selection step of the evolutionary algorithm, see Chapters 22-30). Some
members of the new population undergo transformations by means of genetic
operators to form new solutions. There are unary transformations m, (mutation
type), which create new individuals by a (usually small) change in a single
individual (m, : I -+ I ) , and higher-order transformations c, (crossover, or
recombination type), which create new individuals by combining parts from
several (two or more, up to the population size p ) individuals (cj : I + I ,
2 5 s 5 p).
It seems that, for any evolutionary computation technique, the representation
of an individual in the population and the set of operators used to alter its genetic
code constitute probably the two most important components of the system,
and often determine the systems success or failure. Thus, a representation of
object variables must be chosen along with the consideration of the evolutionary
computation operators which are to be used in the simulation. Clearly, the
reverse is also true: the operators of any evolutionary system must be chosen
carefully in accordance with the selected representation of individuals. Because
of this strong relationship between representations and operators, the latter are
discussed with respect to some (standard) representations.
In general, Chapters 3 1-34 provide a discussion on many operators which
have been developed since the mid-1960s. Chapter 32 deals with mutation
operators. Accordingly, several representations are considered (binary strings,
real-valued vectors, permutations, finite-state machines, parse trees, and others)
and for each representation one or more possible mutation operators are
discussed. Clearly, it is impossible to provide a complete overview of all
mutation operators, since the number of possible representations is unlimited.
TEAM LRN
235
236
References
Eiben A E, Raue P-E and Ruttkay Zs 1994 Genetic algorithms with multi-parcnt
recombination Pro<*.Parcillel Prohleni Sohliiig from Ncitirre vol 3 (New York:
Springer) pp 78-87
Radcliffe N J 1993 Genetic set recombination Foiriicldom of Genetic Algorithms I1
(Ocrober f W 4 , Jerirstilrni) cd Yu Davidor. H-P Schwefel and R Minner (San Matco,
CA: Morgan Kaufmann) pp 203-19
TEAM LRN
32
Mutation operators
Thomas Back (32.1)) David B Fogel (32.2, 32.4, 32.6))
Darrell Whitley (32.3) and Peter J Angeline (32.5, 32.6)
(32.1)
TEAM LRN
237
238
Mutation operators
where 1.1 c / ( [ O , I ) ) denotes a uniform random variable sampled anew for each
i E ( I ,...,t ) .
From a computational point of view, the straightforward implementation
of equation (32.1) as a loop calling the random number generator for each
position i is extremely inefficient. Since the random variable T describing the
distances between two positions to be mutated has a geometrical distribution wi1.h
P { T = t } = p,( I - P,,~)- and expectation E[T] = l / ~ , ? and
~ , a geometrical
random number can be generated according to
In(l - U )
- Pm)
(32.2)
It is obvious from these results that not only for evolution strategies and
evolutionary programming, but also for canonical genetic algorithms, mutation
is an important search operator that cannot be neglected either in practical
applications or in theoretical investigations of these algorithms. Moreover. it
is also possible to release the user of a genetic algorithm from the problem
of finding an appropriate mutation rate control or fine-tuning a fixed value
by transferring the strategy parameter self-adaptation principle from evolution
TEAM LRN
strategies and evolutionary programming
to genetic algorithms.
Real-valued vectors
239
David B Fogel
Mutation generally refers to the creation of a new solution from one and only
one parent (otherwise the creation is referred to as a recombination (see Chapter
33). Given a real-valued representation where each element in a population is an
n-dimensional vector x E R, there are many methods for creating new elements
(offspring) using mutation. These methods have a long history, extending back
at least to Bremermann (1962), Bremermann et a1 ( I965), and others. A variety
of methods will be considered here.
The general form of mutation can be written as
2 = m ( x )
(32.3)
where x is the parent vector, m is the mutation function, and xis the resulting
offspring vector. Although there have been some attempts to include mutation
operators that do not operate on the specific values of the parents but instead
simply choose z from a fixed probability density function (PDF) (Montana and
Davis 1989), such methods lose the inheritance from parent to offspring that
can facilitate evolutionary optimization on a variety of response surfaces. The
more common form of mutation generates an offspring vector:
x=x+M
(32.4)
where the mutation M is a random variable. M is often zero mean such that
E ( z ) = x;the expected difference between a parent and its offspring is zero.
M can take different forms. For example, M could be the uniform random
variable U ( a , b), where a and 6 are the lower and upper limits respectively. In
this case, a is often set equal to -b. The result of applying this operator as M in
equation (32.4) yields an offspring within a hyperbox z U ( - h . h). Although
such a mutation is unbiased with respect to the position of the offspring within
the hyperbox, the method suffers from easy entrapment when the parent vector
x resides in a locally optimal well that is wider than the available step size.
Davis ( 1989, 1991b) offered a similar operator (known as creep) that has a
fixed probability of altering each component of x up or down by a bounded
small random amount. The only method for alleviating entrapment in such cases
relies on probabilistic selection, that is, maintaining a probability for choosing
lesser-valued solutions to become parents of the subsequent generations (see
Chapter 27). In contrast, unbounded mutation operators do not require such
selection methods to guarantee asymptotic global convergence (Fogel 1994.
Rudolph 1994).
The primary unbounded mutation PDF for real-valued vectors has been the
Gaussian (or normal) (Rechenberg 1973, Schwefel 1981, Fogel et a1 1990,
TEAM LRN1993, Fogel and Stayton 1994, and
Fogel and Atmar 1990, Back and Schwefel
240
Mutation operators
When p = 0, the parameter (T offers the single control on the scaling of the
PDF. It effectively generates a typical step size for a mutation. The use of zeromean Gaussian mutations generates offspring that are (i) on average no different
from their parents and (ii) increasingly less likely to be increasingly different
from their parents. Saltations are not completely avoided such that any local
optimum can be escaped from in a single iteration, yet they are not so common
as to lose all inheritance from parent to offspring.
Other density functions with similar characteristics have also been
implemented. Yao and Liu (1996) proposed using Cauchy distributions to aid
in escaping from local minima (the Cauchy distribution has a fatter tail than the
Gaussian) and demonstrated that Cauchy mutations may offer some advantages
across a wide testbed of problems. Montana and Davis (1989) examined the
use of Laplace-distributed mutations but there is no evidence that the Laplace
distribution is particularly better suited than Gaussian or Cauchy mutations for
typical real-valued optimization problems.
In the simplest version of evolution strategies or evolutionary programming.
described as a ( 1 + 1 ) evolutionary algorithm, a single parent x creates a single
offspring xby imposing a multivariate Gaussian perturbation with mean zero
and standard deviation (T on the parent, then selects the better of the two trial
solutions as the parent for the next iteration. The same standard deviation is
applied to each component of the vector x during mutation. For some problems,
the variation of (T (i.e. the step size control parameter in each dimension) c m
be computed to yield an optimal rate of convergence.
Let the convergence rate be defined as the ratio of the Euclidean distance
covered toward the optimum solution to the number of trials required to achieve
the improvement. Rechenberg ( 1973) calculated the convergence rates for two
functions:
(2, . . . , n ] -b/2 5 x l 5 h / 2
24 I
Real-valued vectors
on the corridor model, and
a = 1.22411x11/n
on the sphere model. That is, only a single step size control is needed for
optimum convergence. Given these optimum standard deviations for mutation,
the optimum probabilities of generating a successful mutation can be calculated
as
pyp' = (2e)-I
* 0.184
opt
p 3 = 0.270.
Noting the similarity of these two values, Rechenberg (1973) proposed the
following rule:
The ratio of successful mutations to all muta ions should be l / S . If this rati 1 is
greater than 1/5, increase the variance; if it is less, decrease the variance.
Schwefel (198I ) suggested measuring the success probability on-line over O??
trials (where there are 12 dimensions) and adjusting a at iteration t by
a(t - n)6
a ( t - 17)
if p , < 0.2
if p \ > 0.2
if p\ = 0.2
242
Mutation operators
= U / exp(roN(0, 1 )
x, = x,
+ N(0,
+ riv,(0, 1 ) )
U/)
where the parents strategy parameters are used to create the offsprings
objective values before being mutated themselves, and the mutation of the
strategy parameters is achieved using a Gaussian distribution scaled by x
and the standard deviation for each dimension. This procedure also requires
incorporating a rule such that if any component a/becomes negative it is reset
to an arbitrary small value 6 .
Several comparisons have been conducted between these methods.
Saravanan and Fogel (1994) and Saravanan et a/ (1995) indicated that ihe
log-normal procedure offered by Schwefel ( 198I ) generated generally superior
optimization performance (statistically significant) across a series of standxd
test functions. Angeline (1996a), in contrast, found that the use of Gaussian
mutations on the strategy parameters generated better optimization performarice
when the objective function was made noisy. Gehlhaar and Fogel (1906)
indicated that mutating the strategy parameters before creating the offspring
objective values appears to be more generally useful both in optimizing a set of
test functions and in molecular docking applications.
Both of the above methods for self-adaptation have been extended to
include possible correlation across the dimensions. That is, rather than us(: n
independent Gaussian random perturbations, a multivariate Gaussian mutat ion
with arbitrary covariance can be applied. Schwefel (1981) described a method
for incorporating rotation angles a such that new solutions are created by
Permutations
243
+ A ( t , ub,
-XI@))
if U < 0.5
if u 2 0.5
where x ; ( t ) is the ith parameter of the vector x at generation t , s,E [Ib,. ub,],
the lower and upper bounds, respectively, U is a random uniform U ( 0 , I ) , and
the function A ( ? ,y ) returns a value in the range [0, y ] such that the probability
of A ( r , y) being close to zero increases as t increases, essentially taking smaller
steps on average. Michalewicz et a1 (1994) used the function
A ( r , y ) = yu(1
t/T)"
32.3 Permutations
Darrell Whitley
32.3.I
Introduction
244
Mutation operators
A B C D E F G H
the 2-opt operator selects two points along the string, then reverses the segment
between the points. Note that if the permutation is viewed as a circuit as in the
traveling salesman problem (TSP), then all shifts of a sequence of N elements
are equivalent. It follows that once two cut points have been selected in this
circular string, it does not matter which segment is reversed; the effect is the
same.
The 2-opt operator can be applied to all pairs of edges in N ( N - I )/2 steps.
This is analogous to one iteration of local search over all variables in a parameter
optimization problem. If a full iteration of 2-opt to all pairs of edges fails to
find an improving move, then a local optimum has been reached.
G
Figure 32.1. A graph.
2-opt is classically associated with the Euclidean TSP. Consider the graph
in figure 32.1. If this is interpreted as a Euclidean TSP, then reversing the
segment [C D E F] or the segment [G H A B] results in a graph where none of
the edges cross and which has lower cost than the graph where the edges cross.
Let {A, B, . . . , 2 ) be a set of vertices and (a, b) be the edge between vertices A
and B. If vertices {B, C, F, G} in figure 32.1 are connected by the set of edges
((b,c), (b,f), (b,g), (c,f), (c,g) (f,g)), then two triangles are formed when B
is connected to F and C is connected to G. To illustrate, create a new graph
by placing a new vertex X at the point where the edges (b,f) and (c,g) cross.
In the new graph in Euclidean space, the distance represented by edge ( ( 7 , ~ )
must be less than edges (b, x) (x, c), assuming B, C, and X are not on a line;
likewise, the distance represented by edge (f, g) must be less than edge (f, x )
(x,g). Thus, reversing the segment [C D E F] will always reduce the cost of
the tour due to this triangle inequality. For the TSP this leads to the general
principle that multiple applications of 2-opt will always yield a tour that has no
TEAM LRN
crossed edges.
Permutations
245
One can also look at reversing more than two segments at a time. The
3-opt operator cuts the permutation into three segments and then looks at all
possible ways of reordering these segments. There are 3 ! = 6 ways to order the
segments and each segment can be placed in a forward or reverse order. This
yields up to 23 * 6 = 48 possible new reorderings of the original permutation.
For the symmetric TSP, however, all shifted arrangements of the three segments
are equal and all reversed arrangements of the three segments are equal. Thus,
the 3! orderings are all equivalent. (By analogy, note that there is only one
possible Hamiltonian circuit tour between three cities.) This leaves only 23 = 8
ways of placing each of the segments in a forward or reverse direction, each
of which yields a unique tour. Thus, for the symmetric TSP, the cost to test
one 3-opt move is eight times greater than the cost of testing one 2-opt move.
For other types of scheduling problem, such as resource allocation, reversals
and shifts of the complete permutation are not necessarily equivalent and the
cost of a 3-opt move may be up to 48 times greater than that of a 2-opt move.
Also note that there are
ways to break a permutation up into combinations
ways of breaking the permutation into two
of three segments compared to
segments. Thus, the set of all possible 3-opt moves is much larger than the set
of possible 2-opt moves. This further increases the cost of performing one pass
of 3-opt over all possible ways of partitioning a permutation into three segments
compared to a pass of 2-opt over all pairs of possible segments.
One can also use k-opt, where the permutation is broken into k segments,
but such an operator will obviously be very costly.
(y)
(1)
,
scramble operators
32.3.3 Insert, s ~ u pand
246
Mutation operators
suggest that a single mutation should represent a minimal change and look at
different types of mutation operator for different representations of the TSP.
For resource allocation problems, a more modest change than 2-opi is
to merely select one element and to insert it at some other position in the
permutation. Syswerda (1991) refers to a variant of this as position-based
mutation and describes it as selecting two elements and then moving the second
element before the first element. Position-based mutation appears to be less
general than the insert operator, since elements can only be moved forward in
position-based mutation.
Similarly, one can select two elements and swap the positions of the two
elements. Syswerda denotes this as order-based mutation. Note that if an
element is moved forward or backward one position, this is equivalent to a
swap of adjacent elements. One way in which swap can be used as a local
search operator is to swap all adjacent elements, or perhaps also all pairs of
elements. Finally, Syswerda also defines a scramble mutation operator that
selects a sublist of permutation elements and randomly reorders (i.e. scrambles)
the order of the subset while leaving the other elements in the permutation in
the same absolute position. Davis (1991a) also reports on a scramble sutilist
mutation operator, except that the sublist is explicitly composed of contiguous
elements of a permutation. (It is unclear whether Syswerdas scramble operator
is also meant to work on contiguous elements or not; an operator that selects
a sublist of elements over random positions of the permutation is certainly
possible.)
For a problem that involved scheduling a limited number of flight simulators,
Syswerda ( 199I , p 342) reported that when applied individually, the order-based
swap mutation operator yielded the best results when compared to positionbased mutation and scramble mutation. In this case the swaps were selected
randomly rather than being performed over a fixed well-defined neighborhood.
Davis ( 1991, p 8 I ) on the other hand reports that the scramble sublist mutation
operator proved to be better than the swap operator on a number of applications.
In conclusion, one cannot make a priori statements about the usefulness of
a particular mutation operator without knowing something about the type of
problem that is to be solved and the representation that is being used for that
problem, but in general it is useful to distinguish between permutation problems
that are sensitive to adjacency (e.g. the TSP) versus relative order (e.g. resource
scheduling) or absolute position, which appears to be the least common.
M =TEAM
((3, TLRN
, P , S, n)
Finite-state machines
247
where Q is a finite set, the set of states, T is a finite set, the set of input
symbols, P is a finite set, the set of output symbols, s : Q x T -+ Q , the next
state function, and o : Q x T -+ P , the next output function,
there are various methods for mutating parents to create offspring. Following
directly from the definition, five obvious modes of mutation present themselves:
(i) change an output symbol, (ii) change a state transition, (iii) add a new state,
(iv) delete a state, and (v) change the start state. Each of these will be discussed
in turn.
(i)
(ii)
(iii)
(iv)
(v)
248
Mutation operators
Parse trees
249
Figure 32.2. An illustration of the grow mutation operator applied to a Boolean parse
tree. Given a parent tree to mutate, a terminal node is selected at random (highlighted)
and replaced by a randomly generated subtree to produce the child tree.
Figure 32.3. An illustration of the shrink mutation operator applied to a Boolean parse
tree. Given a parent tree to mutate, an internal function node is selected at random
(highlighted) and replaced by a randomly selected terminal to produce the child tree.
Figure 32.4. An illustration of the switch mutation operator applied to a Boolean parse
tree. Given a parent tree to mutate, an internal function node is selected, two of the
subtrees below it are selected (highlighted in the figure) and their positions switched to
produce the child tree.
TEAM LRN
250
Mutation operators
Figure 32.5. An illustration of the cycle mutation operator applied to a Boolean parse
tree. Given a parent tree to mutate, a single node, either a terminal or function, is selected
at random (highlighted in the parent) and replaced by a randomly selected node with the
same number of arguments to produce the child tree.
Many real-world applications suggest the use of representations that are hybrids
of the canonical representations. One common instance is the simultaneous use
of discrete and continuous object variables, with a general formulation of the
global optimization problem as follows (Back and Schiitz 1995):
min{f(x, d ) ( x E M , R" 2 M , d
N , 2"" 1N } .
Other representations
25 I
Back and Schutz (1995) approach the general problem by including a vector
of mutation strategy parameters p j E (0, 1) and j = I , 2, . . . , d , where there are
d components to the vector d . (Alternatively, fewer strategy parameters could
be used.) These strategy parameters are adapted along with the usual step size
control strategy parameters for Gaussian mutation of the real-world vector x.
The discrete strategy parameters are updated by the formula
252
Mutation operators
References
Angeline P J I996a The effects of noise on self-adaptive evolutionary optimization, Pmc.
5th Ann. Cot$ oti E\lolittioncin Progrunttnirtg ed L J Fogel, P J Angeline and T
Back (Cambridge, MA: MIT Press) pp 443-50
-1996b Genetic programmings continued evolution Adiutices in Genetic Progr,itnming vol 2, ed P J Angeline and K Kinnear (Cambridge, MA: MIT Press) pp 89-1 10
Angeline P J, Fogel D B and Fogel L J 1996 A comparison of self-adaptation methods for
finite state machines in a dynamic environment Proc. 5th Atin. Cot$ otz Evo1utior:cin.
Progrcim?ting ed L J Fogel, P J Angeline and T Back (Cambridge, MA: MIT Przss)
pp 431-50
Back T 1993 Optimal mutation rates in genetic search Proc. 5th Itit. Cot$ oti Genetic
Algorithms (UrbcitiCi-Chat)ipcrign. IL, July 1993) ed S Forrest (San Mateo, CA:
Morgan Kaufmann) pp 2-8
-1996 E\~olutiottciry Algorithnrs in Theory und Prcicticbe (New York: Oxford
University Press)
Biick T, Rudolph G and Schwefel H-P 1993 Evolutionary programming and evolution
strategies: similarities and differences Proc. 2nd Anti. Cot$ on E\dutiorrcin
Progranimi,ig (Sun Diego, CA) ed D B Fogel and W Atmar (La Jolla, CA:
Evolutionary Programming Society) pp 1 1-22
Back T and SchiitL M 1995 Evolution strategies for mixed-integer optimization of optical
i
Progrummirig (San Diego,
multilayer systems Proc. 4th Anri. Cortf. o ~ Evolittioncin
CA, March 1995) ed J R McDonnell, R G Reynolds and D B Fogel (Cambridge.
MA: MIT Press) pp 33-51
Biick T and Schwefel H-P 1993 An overview of evolutionary algorithms for parameter
optimization E\.wlutiotui n-Cotnput. 1 1-24
Bagley J D 1967 The BehuiGor of A d a p t i , ~SJAtetns \t*hich Employ Genetic cind
Correlcitioti Algorithnis Doctoral Dissertation, University of Michigan; Universitj
Microfilms 68-7556
Bremermann H J 1962 Optimidation through evolution and recombination SelfOrglinizing Sy.\tems ed M C Yovits, G T Jacobi and G D Goldstine (Washington,
DC: Spartan) pp 93-106
Bremermann H J and Rogson M 1964 An Evolution -type Secirch Method j i w Corii-e.x Sets
ONR Technical Report, contracts 222( 85) and 3656(58)
Bremermann H J, Rogson M and Salaff S 1965 Search by evolution BiophyAict arid
Cjhernetic System.\ ed M Maxtield, A Callahan and L J Fogel (Washington, DC:
TEAM LRN
Spartan) pp 157-67
References
25 3
254
Mutation operators
References
255
(Pittsburgh, PA, July 1995) ed L J Eshelman (San Mateo, CA: Morgan Kaufmann)
pp 159-66
Ostermeier A 1992 An evolution strategy with momentum adaptation of the random
number distribution Parallel Problem Solving from Nature, 2 (Proc. 2nd Int. Cortf:
on Parallel Problem Solving from Nature, Brussels, 1992) ed R Manner and B
Manderick pp 197-206
Radcliffe N and Surry P D 1995 Fitness variance of formae and performance prediction
Foundations of Genetic Algorithms 3 ed D Whitley and M Vose (San Mateo, CA:
Morgan Kaufmann) pp 5 1-72
Rechenberg I 1973 Evolutionsstrategie: Optimierung rechrtischer Systeme nach
Prinzipien der biologischen Evolution (Stuttgart: Frommann-Holzboog)
Reed J, Toombs R and Barricelli N A 1967 Simulation of biological evolution and
machine learning J. Theor. Biol. 17 3 1 9 4 2
Rudolph G I994 Convergence properties of canonical genetic algorithms IEEE Tram.
Neural Nemorks 5 96-1 01
Saravanan N and Fogel D B 1994 Learning strategy parameters in evolutionary
programming: an empirical study Proc. 3rd Ann. Conf on Eivlutionap
Programming (San Diego, CA, February 1994) ed A V Sebald and L J Fogel
(Singapore: World Scientific) pp 269-80
Saravanan N, Fogel D B and Nelson K M 1995 A comparison of methods for selfadaptation in evolutionary algorithms BioSystems 36 157-66
Schwefel H-P 1981 Numerical Optimization of Computer Models (Chichester: Wiley)
-1995
Evolution and Optimum Seeking (New York: Wiley)
Schaffer J D, Caruana R A, Eshelman L J and Das R 1989 A study of control parameters
affecting online performance of genetic algorithms for function optimization Proc.
3rd Int. Con$ on Genetic Algorithms (Faigm, VA, June 1989) ed J D Schaffer (San
Mateo, CA: Morgan Kaufmann) pp 51-60
Syswerda G 199 1 Schedule optimization using genetic algorithms Handbook qf Genetic
Algorithms ed L Davis (New York: Van Nostrand Reinhold) pp 332-49
Yao X and Liu Y 1996 Fast evolutionary programming Proc. 5th Ann. Con$ on
Evolutionary Programming ed L J Fogel, P J Angeline and T Back (Cambridge,
MA: MIT Press) at press
TEAM LRN
33
Recombination
Lashon B Booker (33. I), David B Fogel (33.2, 33.4, 33.6 I,
Darrell Whitley ( 3 3 3 , Peter J Angeline (33.5, 33.6)
a n d A E Eiben (33.7)
33. I . I
Introduction
256
TEAM LRN
Binary strings
257
Two new resultant strings are formed by exchanging the parent substrings to the
right of position k . Holland points out that when the overall algorithm is limited
to producing only one new individual per generation, one of the resultant strings
generated by this crossover operator must be discarded. The discarded string is
usually chosen at random.
Holland's general procedure defines a family of operators that can be
described more formally as follows. Given a space I of individual strings,
a crossover operator is a mapping
where m
Bt and
if m , = 0
if m i = 1
di =
{"
a,
if m , = 0
if i n, = I .
258
Recombination
:= bi;
n; :=
a;;
fi
od
return(c, d ) ;
Empirical studies have shown that the best setting for the crossover rate pc
depends on the choices made regarding other aspects of the overall algorithm,
such as the settings for other parameters such its population size and mutation
rate, and the selection operator used. Some commonly used crossover rates
are p c = 0.6 (De Jong 1975), p c E [0.45,0.95] (Grefenstette 1986), and
p , E [0.75,0.95](Schaffer et nl 1989). Techniques for adaptively modifying the
crossover rate have also proven to be useful (Booker 1987, Davis 1989, Srinivas
and Patnaik 1994, Julstrom 1995). The pseudocode shown above makes it clear
that the differences between crossover operators are most likely to be found in
the implementation of the compute-mask() procedure. The following examples
of pseudocode characterize the way compute-mask() is implemented for the
most commonly cited crossover operators.
One-point crossover. A single crossover point is selected. This operator can
only exchange contiguous substrings that begin or end at the endpoints of the
chromosome. This is rarely used in practice.
sample U E U ( 1 , t! - 1)
m := 0;
fori:=u+l told0
m ,= 1 ;
od
return m;
TEAM LRN
Binary strings
259
U,,
fi
m := 0;
for j := 1 to n step 2 do
for i := U, 1 to u,+l do
m ,= 1 ;
od
od
return m;
m := 0;
for i := 1 to l do
sample U E U ( 0 , 1 )
if (U 5 P d
then m , = 1 ;
fi
od
return m
TEAM LRN
260
Recombination
The value p , = 0.5 first used by Ackley remains the standard setting for
the crossover probability at each position, though it may be advantageous to
use smaller values (Spears and De Jong 1991b). When p r = 0.5, every binary
string of length e is equally likely to be generated as a mask. In this case,
i t is often more efficient to implement the operator by using a random integer
sampled from U ( 0 ,2' - I ) as the mask instead of constructing the mask one bit
at a time.
fi
od
return ( m ) ;
Note that the symbol and punctuation mark associated with a chromosome
position are transmitted together by the punctuated crossover operator. While
the idea behind this operator is appealing, empirical tests of punctuated crossover
were not conclusive and the operator is not widely used.
Binary strings
26 I
offspring can substantially reduce the loss of diversity in the population. Another
widespread practice is to restrict the crossover points to those locations where
the parent strings have different symbols. This so-called reduced surrogate
technique (Booker 1987) improves the ability of crossover to produce offspring
that are different from their parents.
An implementation technique called shuffle crosso\er was introduced by
Eshelman et a1 (1989). The symbols in the parent strings are shuffled by a
permutation operator before crossover is invoked. The inverse permutation is
applied to the offspring produced by crossover to restore the original symbol
ordering. This method can be used to counteract the tendency in n-point
crossover ( I ? 2 1) to disrupt sets of symbols that are widely dispersed on the
chromosome more than it disrupts symbols which are close together (see the
discussion of bias in Section 33.1.4).
The crossover mechanisms described so far are all consistent with the
simplest principle of Mendelian inheritance: the requirement that every gene
carried by an offspring is a copy of a gene inherited from one of its parents.
Radcliffe ( 1991 ) points out that this conservation of genetic material during
recombination is not a necessary restriction for artificial recombination operators.
From the standpoint of conducting a robust exploration of the opportunities
represented by the parent strings, it is reasonable to ask whether a crossover
operator can generate all possible offspring having some combination of genes
found in the parents. Given a binary string representation, the answer for onepoint and n-point crossover is no while the answer for shuffle crossover and
uniform crossover is yes. (To see this, simply consider the set of possible
resultant strings for the parents 0 and 1.) For nonbinary strings, however, the
only way to achieve this capability is to allow the offspring to have genes
that are not carried by either parent. Radcliffe used this idea as the basis for
designing the random respectful recombination operator. This operator generates
a resultant string by copying the symbols at positions where the parents are
identical, then choosing random values to fill the remaining positions. Note that
for binary strings, random respectful recombination is equivalent to uniform
crossover with p , = 0.5.
Recombination
262
xAcS,
Binary strings
263
P ~ ' ( Z )=
A.
21
R(A)p"'(z)p")(y)
(33.1)
A.x,y
SE is arbitrary
xi = z; Vi E A'
B EA
(33.3)
These recurrence relations are equivalent, complete characterizations of how
recombination changes the proportion of individuals from one generation to the
next. Equation (33.1) has the straightforward interpretation that alleles appear
in offspring if and only if they appear in the parents and are transmitted by
a recombination event. Each term on the right-hand side of (33.1) is the
probability of a recombination event between parents having the desired alleles at
the loci that are transmitted together. A string z is the result of a recombination
event A whenever the alleles of z at loci A come from one parent and the
alleles at loci A' come from the other parent. The change in frequency of an
individual string is therefore given by the total probability of all these favorable
occurrences. Equation (33.2) is derived from (33.1) by collecting terms based
on marginal recombination probabilities. Equation (33.3) is derived from (33.1)
by collecting terms based on marginal frequencies of individuals.
The last equation is perhaps the most significant, since it leads directly to a
theorem characterizing the expected distribution of individuals in the limit.
Theorem (Geiringer's theorem U). If t loci are arbitrarily linked, with the one
exception of 'complete linkage', the distribution of transmitted alleles 'converges
toward independence'. The limit distribution is given by
TEAM LRN
264
Recombination
which is the product of the t marginal distributions of alleles from the initial
popu M o n .
This theorem tells us that, in the limit, random mating and recombination
without selection lead to chromosome frequencies corresponding to the simple
product of initial allele frequencies. A population in this state is said to be
in lirzkcige eqiiilibriiirn or Robbins equilibrium (Robbins 19 18). This result
holds for all recombination operators that allow any two loci to be separated by
recombination.
Note that Holland (1975) sketched a proof of a similar result for schetna
frequencies and one-point crossover. Geiringers theorem applied to schemata
gives us a much more general result. Together with the recurrence equations,
this work paints a picture of search pressure from recombination acting to
reduce departures from linkage equilibrium for all schemata.
Subsequent work has carefully analyzed the dynamics of this convergence
to linkage equilibrium (Christiansen 1989). It has been proven, for example,
that the convergence rate for any particular schema is given by the probability
of the recombination event specified by the schemas defining loci. In this
view, an important difference between crossover operators is the rate at which,
undisturbed by selective pressures, they drive schemata to their equilibrium
proportions. These results from mathematical population genetics have oiily
recently been applied to evolutionary algorithms (Booker 1993, Altenberg 1995).
265
Binary strings
the loci in A . The index set expression
makes it clear that p A ( B ) involves strings having the allele values given by B
at the loci designated by A . Note that p f l ( B )= 1 and p s , ( B ) = p ( B ) .
With this notation we can also succinctly relate recombination distributions
A
and schemata. If A designates the defining loci of a schema and B
specifies the alleles at those loci, then the frequency of 6 is given by p , 4 ( B )and
the marginal distribution RAdescribes the transmission of the defining loci of
6. In what follows we will assume, without loss of generality, that the elements
of the index set A for a schema 6 are in increasing order so that the kth element
A ( A )is the locus of the kth defining position of
This means, in particular,
that the outermost defining loci of are given by the elements A " , and A , Q ~ , ,
where O(6) is the order of 6. It will be convenient to define the following
property relating the order of a schema to its defining length a(<).
c.
Dejinition. The kth cwnporient of defining length for schema 6,&, (<), is the
1st defining loci, 1 5 k < O((), with the
distance between the kth and k
convention that
t - S(6).
&(c)
Note that the defining length of a schema is equal to the sum of its defining
length components:
k= 1
RI(&.)
= RI($)=
2(t - I )
since each parent is equally likely to transmit the indicated loci. The marginal
distribution 72; for an arbitrary index set A can be expressed solely in terms
of these recombination events. We will refer to these events as the primary
recombination events.
Now for any arbitrary event B 2 A there are two cases to consider:
(i)
Recomb i n ati 02
266
(ii) B # 8. These situations involve the primary events S,r, A(1) 5 x <
A(o(6)). The events B having nonzero probability are given by B, =
{ A ( 1 ) ,. . . , A , , ) } , 1 5 i < O(6). For each i , there are 6,(6) corresponding
primary events.
The complete marginal distribution is therefore given by
RA(B)=
e - 1 - 6(6)
2 ( [ - 1)
W)
2(e - 1)
0
if B = 8 or B = A
if B = B , , 1 5 i < O(c)
otherwise.
if n is even and { A , , , ,A ( , - I ) }n B = !d or { A ( , ) ,A ( , - I ) )
if n is odd and [ A ( , ) ,A ( [ - l ) } n B # yl or { A ( , ) A
, (,- I)
where 2 5 i 5 O(6)
otherwise. TEAM LRN
Binary strings
267
Putting all the pieces together, we can now give an expression for the
complete marginal distribution.
Figure 33.1 shows how the marginal probability of transmission for secondorder schemata-2 R: ( A ) and 2 R:(o.,
I A J = 2-varies
as a function of
defining length. The shape of the curves depends on whether n is odd or
even. Since the curves indicate the probability of transmitting schemata, the
area above each curve can be interpreted as a measure of potential schema
disruption. This interpretation makes it clear that two-point crossover is the best
choice for minimizing disruption. Spears and De Jong (1991a) have shown that
this property of two-point crossover remains valid for higher-order schemata.
Note that these curves are not identical to the family of curves for
nondisruptive crossovers given by Spears and De Jong. The difference is
TEAM LRNpoints are selected randomly with
that Spears and De Jong assume crossover
268
Recombination
Figure 33.1. Transmission probabilities for second-order schemata. The inset shows the
behavior of these curves in the vicinity of the point L / 2 .
269
Binary strings
search is the only search technique that has no bias. It has long been recognized
that an appropriate inductive bias is necessary in order for inductive search to
proceed efficiently and effectively (Mitchell 1980). Two types of bias have
been attributed to crossover operators in genetic search: di.rtributinnn1 h i n s and
positiorzal bins (Eshelman et a1 1989).
Distributional bias refers to the number of symbols transmitted during a
recombination event and the extent to which some quantities might be more
likely to occur than others. This bias is significant because i t is correlated with
the potential number of schemata from each parent that can be recombined by
the crossover operator. An operator has distributional bias if the probability
distribution for the number of symbols transmitted from a parent is not uniform.
Both one-point and two-point crossover are free of distributional bias. The
n-point ( 1 2 > 2) crossover operators have a distributional bias that is well
approximated by a binomial distribution with mean t / 2 for large i i . Uniform
crossover has a strong distributional bias, with the expected number of symbols
transmitted given by a binomial distribution with expected value p , t . More
recently, Eshelman and Schaffer ( I 993) have emphasized the expected value of
the number of symbols transmitted rather than the distribution of those numbers.
The bias defined by this criterion, though clearly similar to distributional bias,
is referred to as recmnbinati\,e bias.
Positional bias characterizes how much the probability that a set of symbols
will be transmitted intact during a recombination event depends on the relative
2 L '
01 pt
1.5
v)
'
'
'
'
:1 381s
03pts
02pts
04pt95pts
013pts
14pts
1,
-a..
1/2
0.5
ogpts
7pts 1opts
; ;0f"
,
lf8
1.5
Distributional Bias
Figure 33.2. One view of the crossover bias 'landscape' generated using quantitative
TEAM LRN
measures derived from recombination distributions.
270
Recombination
27 1
Real-valued vectors
Parents
21
= Xl,lXl,2
* *
*Xl,kXl,k+l
2 2
= x 2 . Ix2.2
* *
X2.kX2.k+ I
* *
*Xl.tl
* * *
x2.d
Figure 33.3. For one-point crossover, two parents are chosen and a crossover point k
is selected, typically uniformly across the components. Two offspring are created by
interchanging the segments of the parents that occur from the crossover point to the ends
of the string.
ly
Offspring
Figure 33.5. For two-point crossover, two parents are chosen and two crossover points,
k l and k 2 , are selected, typically uniformly across the components. Two offspring are
created by interchanging the \egments defined by the points kl and kz.
Figure 33.6. For uniform crossover, each element of an offspring (here two offspring
art' depicted) is selected from either parent. The example shows that the first element in
both offspring were selected from the first parent. In some applications such duplication
is not allowed. Typically each parent has an equal chance of contributing each elemcnt
to an offspring.
Offspring
Parent 1
Parent 2
on two parents
21
and
22. and
= aI.rIl TEAM
a?.r?,
LRN . . .
+ a/(s/(r
Real-valued vectors
subject to
273
E./ = I
i =l,...,k
where ii is a uniform random variable over [0, 11 and 21 and 2 2 are the
two parent vectors subject to the condition that 21 is not worse than 21.
Michalewicz (1996, p 129) noted that this operator uses values of the
objective function to determine a direction to search.
(ii) The simplex crossover of Renders and Bersini ( 1994) selects k > 2 parents
(say the set J ) , determines the best and worst individuals within the selected
group (say 21 and 22, respectively), computes the centroid of the group
without 2 2 (say c ) and computes the reflected vector 2' (the offspring)
obtained from the vector 22 as
2' = c
+ (c
- 2,).
2' = [ ( X I lxzl)o.s,. . , ( - T 1 , / S 2 , 1 ) o . s ] .
273
Recombination
(iv) The jitness-based scan of Eiben et a1 (1904) takes multiple parents arid
generates an offspring where each component is selected from one of the
parents with a probability corresponding to the parents relative fitness. If
a parent has fitness f ( i ) , then the likelihood of selecting each individual
component from that parent is f ( i ) /
f ( j ) ,where j = I , . . . , k and there
are k parents involved in the operator.
( v ) The diagonal multiparent crossover of Eiben et a1 (1994) operates much
like n-point crossover, except that in creating k offspring from k parents,
c >_ 1 crossover points are chosen and the first offspring is constructed to
contain the first segment from parent 1 , the second segment from parent 2,
and so forth. Subsequent offspring are similarly constructed from a rotation
of segments from the parents.
33.3 Permutations
I Whit1ey
Dli rre 1
33.3.I
Iritroduc*tion
crossover operators fail to generate offspring that are permutations. Consider the
following example of simple one-point crossover, when one parent is denoted
with capital letters and the other with lower-case letters:
String 1: A B C D E F G H I
\/
/\
String 2: h d a e i c f b g
Offspriqg 1: A B C e 1 c f b g
Offspring 2: h d a D E F G H I.
Permutations
275
from the filler string into the next available slot in the offspring. Continue
moving the next unused element from the filler string to the offspring. When
the end of the filler string (or the offspring) is reached, wrap around to the
beginning of the string. When done in this way, Daviss order crossover has
the property that Radcliffe (1 994) describes as pure recombination: when two
identical parents are recombined the offspring will also be identical with the
parents. If one does not start copying elements from the filler string starting at
the second crossover point, the recombination may not be pure.
The following is an example of Daviss order crossover, where dots
represent the crossover points. The underscore symbol in the crossover section
corresponds to empty slots in the offspring.
Parent 1: A B . C D E F . G H I
Crossover-section: - - C D E F - - Parent 2: h d . a e i c . f b g
Available elements in order: b g h a i
Offspring: a i C D E F b g h.
Note that the elements in the crossover section preserve relative order,
absolute position, and adjacency from parent 1 . The elements that are copied
from the filler string preserve only the relative order information from the second
parent.
Partially mapped crossover (PMX). Goldberg and Lingle ( 1985) introduced the
partially mapped crossover operator (PMX). PMX shares the following attributes
with Daviss order crossover. One parent string is designated as parent 1, the
other as parent 2. Two crossover sites are selected and all of the elements in
parent 1 between the crossover sites are directly copied to the offspring. This
means that PMX also defines a crossover section in the same manner as order
crossover.
Parent 1: A B . C D E . F G
Crossover-section: - - C D E Parent 2: c f
e b a
d g.
The difference between the two operators is in how PMX copies elements
from parent 2 into the open slots in the offspring after a crossover section has
been defined. Denote the parents as P1 and P2 and the offspring as OS; let
P1, denote the ith element of permutation PI. The following description of
selecting elements from P2 to place in the offspring is based on the article by
Whitley and Yoo (1995).
For those elements between the crossover points in parent 2, if element P2,
has already been copied to the offspring, take no action. In the example given
here, element e in parent 2 requires no processing. We will consider the rest of
TEAM LRN
the elements by considering the positions
in which they appear in the crossover
276
RecombinatitE
section. If the next element at position P2, in parent 2 has not already bezn
copied to the offspring, then find PI; = P2j; if position j has not been filled
in the offspring then assign OSj = P2;. In the example given here, the next
element in the crossover section of parent 2 is b which is in the same position as
D in parent 1 . Element D is located in parent 2 with index 6 and the offspring at
OS6 has not been filled. Copy h to the offspring in the corresponding position.
This yields
Offspring: - - C D E b
_.
2, but position 3 has already been filled in the offspring. The position in the
offspring is filled by C, so we now find element C in parent 2. The position
is unoccupied in the offspring, so element A is placed in the offspring at the
position occupied by C in parent 2. This yields
Offspring: a - C D E b _ .
All of the elements in parent 1 and parent 2 that fall within the crossover
section have now been placed in the offspring. The remaining elements can be
placed by directly copying their positions from parent 2. This yields
Offspring: a f C D E b g.
The selected elements in parent 2 are F, B, and A. Thus, the relevant elements
are reordered in parent 1.
Reorder A B - - - F - from parent 1 which yields f b - -
a _.
->
( - LRN
- C D E - C> yields f b C D E a G.
combined with
TEAM
277
Permutations
Next scan parent 2 from left to right and place place each element which
does not yet appear in the offspring in the next available position. This yields
the following progression:
# # C D E # G = > f # C D E # G
= > f b C D E # G
= > f b C D E a G .
Obviously, in this case the two operators generate exactly the same offspring.
Jim Van Zant first pointed out the similarity of these two operators in the
electronic newsgroup The Genetic Algorithm Digest. Whitley and Yoo ( 1995)
show the two operators to be identical using the following argument.
Assume there is one way to produce a target string S by recombining two
parents. Given a pair of strings which can be recombined to produce string S. the
probability of selecting the K key positions using order crossover-2 required to
generate a specific string S is (:)-, while for position crossover the probability
of picking the L - K key elements that will produce exactly the same effect is
( L f K ) - l . Since
= ( L f K ) the probabilities are identical.
Now assume there are R unique ways to recombine two strings to generate a
target string S. The probabilities for each unique recombination event are equal
as shown by the argument in the preceding paragraph. Thus the sum of the
probabilities for the various ways of ways of generating S are equivalent for
order crossover-2 and position crossover. Since the probabilities of generating
any string S are identical, the operators are identical in expectation.
This also means that in practice there is no difference between using order
crossover-2 and position crossover as long as the parameters of the operators
are adjusted to reflect their complementary nature. If position crossover is used
so that X% of the positions are initially copied to the offspring. then order
crossover is identical if (100 - A)% positions are selected as relative order
positions.
(k)
278
Recombination
Pick a random city as the initial current city. Remove all references to this
city from the edge table.
(ii) Look at the adjacency list of the current city. If there is a common edge
(flagged by +), go to that city next. (Unless the initial city is the current
TEAM LRN
city, there can be only one common
edge; if two common edges existed,
279
Permutations
+j, m, i
1 I
+d, f, h
one was used to reach the current city.) Otherwise from the cities on the
current adjacency list pick the next city to be the one whose own adjacency
list is shortest. Ties are broken randomly. Once a city is visited, references
to the city are removed from the adjacency list of other cities and it is no
longer reachable from other cities.
(iii) Repeat step 2 until the tour is complete or a city has been reached that has
no entries in its adjacency list. If not all cities have been visited, randomly
pick a new city to start a new partial tour.
Jsing the edge table in figure 33.8, city a is randomly chosen as the first
city in the tour. City k is chosen as the second city in the tour since the edge
(a,k) occurs in both parent tours. City e is chosen from the edge list of city k
as the next city in the tour since this is the only city remaining in ks edge list.
This procedure is repeated until the partial tour contains the sequence [a k e c].
At this point there is no deterministic choice for the fifth city in the
tour. City c has edges to cities d and g, which both have two unused edges
remaining. Therefore city d is randomly chosen to continue the tour. The
normal deterministic construction of the tour then continues until position 7. At
position 7 another random choice is made between cities f and h. City h is
selected and the normal deterministic construction continues until we arrive at
the following partial tour: [a k e c d m h b g].
In this situation, a failure occurs since there are no edges remaining in the
edge list for city g. When a potential failure occurs during edge-3 recombination,
we attempt to continue construction at a previously unexplored terminal point
in the tour.
A terminal is a city which occurs at either end of a partial tour, where all
edges in the partial tour are inherited from the parents. The terminal is said to
be live if that city still has entries in its edge list; otherwise it is said to be a dead
terminal. Because city a was randomly chosen to start the tour in the previous
example, it serves as a new terminal in the event of a failure. Conceptually this
is the same as inverting the partial tour to build from the other end.
When a failure occurs, there is at most one live terminal in reserve at the
opposite end of the current partial tour. In fact, it is not guaranteed to be live,
since the construction of the partialTEAM
tour LRN
could isolate this terminal city. Once
280
Recombination
both terminals of the current partial tour are found to be dead, a new partial
tour must be initiated. Note that no local information is employed.
We now continue construction of the partial tour [a k e c d m h b g]. The
tour segment is reversed (i.e. [g b h m d c e k a]). Then city i is added to the
tour after city a. The tour is then constructed in the normal fashion. In this
case. there are no further failures. The final offspring tour is [g b h m d c e k
a i f j]. The offspring produced has a single foreign edge (&-g].)
When a failure occurs at both ends of the subtour, edge-3 recombination
starts a new partial tour. However, there is one other possibility, which has
been described as part of the edge-4 operator (Dzubera and Whitley 1994) but
which has not been widely tested.
Assume that the first partial tour has been constructed such that both ends
of the construction lack a liile terminal by which to continue. Since only one
partial tour has been constructed and since initially every city has at least two
edges in the edge table, there must be edges internal to the current partial tour
that represent possible edges to the terminal cities of the partial tour. The edge-4
operator attempts to exploit this fact by inverting part of the partial tour so that a
terminal city is reconnected to an edge which is both internal to the partial tour
and which appeared in the original edge list of the terminal city. This will cause
a previously visited city in the partial tour to move to a terminal position. If this
newly created terminal has cities remaining in its (old) edge list, the offspring
construction can continue. If it does not, one can look for other internal edges
that will allow an inversion. Details on the edge-4 recornbination operator are
given by Dzubera and Whitley (1994).
If one is using just a recombination operator and a mutation operator,
then edge recombination works very well as an operator for the TSP, at least
compared to other recombination operators, but if one is hybridizing such that
tours are being produced by recombination, then improved using 2-opt, then
both the empirical and the theoretical evidence suggests that Muhlenbeins MPX
operator may be more effective (Dzubera and Whitley 1994).
Permutations
28 1
(a) IF an edge from the receiver parent starting at the last city in the
offspring is possible (does not violate a valid tour)
(b) THEN add this edge from the receiver
(c) ELSE IF an edge from the donor starting at the last city in the
offspring is possible
(d) THEN add this edge from the donor
(e) ELSE add that city from the receiver which comes next in the string;
this adds a new edge, which we will mark as an implicit mutation.
The following example illustrates the MPX operator.
Receiver:
Donor :
Initial segment:
G D M H B J F I A K E C
c e k a g b h i j f m d
- - k a g - - - - - - _.
Note that MPX does not transmit adjacency information from parents to
offspring as effectively as the various edge recombination operators, since it
uses less lookahead to avoid a break in the tour construction. At the same time,
when it must introduce a new edge that does not appear in either parent, it skips
to a nearby city in the tour rather than picking a random edge. Assuming that
the tour is partially optimized (for example, if the tour has been improved via
2-opt) then a city nearby in the tour should also be a city nearby in Euclidean
space. This, coupled with the fact that an initial segment is copied from one of
the parents, appears to give MPX an advantage when when combined with an
operator such as 2-opt. Gorges-Schleuter ( 1989) implemented a variant of MPX
that has some notable features that are somewhat like Davis's order crossover
operator. A full description of Gorges-Schleuter's MPX is given by Dzubera
and Whitley (1994).
282
Recombination
two parent strings. Consider the following example from Oliver et a1 (1987)
where the permutation elements correspond to the alphabetic characters with
numbers to indicate position:
Parentl: h k c e f d b l a i g j
Parent2: a b c d e f g h i j k 1
P o s i t i o n s : 1 2 3 4 5 6 7 8 9 10 11 12.
To find a cycle, pick a position from either parent. Starting with position 1,
elements (h, a) belong to cycle 1 . The elements (h, a) also appear in positions
8 and 9. Thus the cycle is expanded to include positions ( 1 , 8, 9) and the new
elements i and 1 are added to the corresponding subset. Elements i and I appear
in positions 10 and 12, which also causes j to be added to the subset of elements
in the cycle. Note that adding j adds no new elements, so the cycle terminates.
Cycle I includes elements (h, a, i, I, j ) in positions ( I , 8, 9, 10, 12).
Note that element (c) in position 3 forms a unary cycle of one element.
Aside from the unary cycle at element c (denoted U), Oliver et a1 note that
there are three cycles between this set of parents:
Parentl:
h k c e f d b l a i g j
Parent2:
a b c d e f g h i j k l
Cycle Label: 1 2 U 3 3 3 2 1 1 1 2 1.
Recombination can occur by picking some cycles from one parent and
the remaining cycles from the alternate parent. Note that all elements in the
offspring occupy the same positions as in one of the two parents. However, few
applications seem to be position sensitive and cycle crossover is less effective at
preserving adjacency information (as in the TSP) or relative order information
(as in resource scheduling) compared to other operators.
Permutations
283
placed in the offspring. Because C has already been allocated a position in the
offspring, the C which appears later in parent 2 is exchanged with the E in the
initial position of parent 2. This yields
Parent 1:
C F G B A H D I E J
Parent 2 :
C B G J D I <E> A F H
Precedence: A B C D E F G H I J
I E J
<E> A F H
H I J.
Note that one need not actually build a separate offspring, since both parents
are in effect transformed into copies of the same offspring. The resulting
offspring in the above example is
Offspring: C B G F A H D E I J .
The MX-2 operator is similar, except that when an element is added to the
offspring it is deleted from both parents instead of being swapped. Thus, the
process works as follows:
Parent 1:
Parent 2 :
Precedence:
C F G B A H D I E J
E B G J D I C A F H
A B C D E F G H I J.
D I E J
A F H
Instead of now moving to the second element of each permutation, the first
remaining elements in the parents are compared: in this case, E and F are the
first elements and E is chosen and deleted. The parents are now represented as
follows:
Parent 1: - F G B A H
Parent 2 : - B G J D I
Offspring: C E.
D I
A F H
Note that, over time, this class of operator will produce offspring that are
closer to the precedence vectordven
if LRN
no selection is applied.
TEAM
284
Recombination
Other interesting operators have been introduced over the years for permutation
problems. Fox and McMahon (1991) introduced an intersection operator that
extracts features common to both parents. Eshelman (1991) used a similar
strategy to build a recombination operator that extracts all common subtours
for the TSP, and assigns all other elements using local search (2-opt) over an
otherwise random assignment. Fox and McMahon also constructed a union
operator. In this case, each permutation is converted into a binary matrix
representation and the offspring is the logical-or of the matrices representing
the parents.
Radcliffe and Surry (1995) have also introduced new operators for the TSP,
largely by looking at different representations and then defining appropriate
operators with respect to the representations. These representations include the
permutation representation, the undirected edge representation, the directed edge
representation, and the corner representation.
M = ( Q , T , P , S , 01
where Q is a finite set, the set of states, T is a finite set, the set of input
symbols, P is a finite set, the set of output symbols, s : Q x T -+ Q, the
next state function, and o : Q x T --+ P , the next output function. Perhaps the
earliest proposal to recombine finite-state machines in simulated evolution can
be found in the work of Fogel (1964) and Fogel et a/ (1966, pp 21-3). The
following extended quotation (Fogel et a1 1966, p 21) may be insightful:
The recombination of individuals of opposite sex appears to benefit
natural evolution. By analogy, why not retain worthwhile traits that
have survived separate evolution by combining the best surviving
machines through some genetic rule; mutating the product to yield
offspring? Note that there is no need to restrict this mating to the
best two surviving individuals. In fact, the most obvious genetic
rule, majority logic, only becomes meaningful with the combination
of more than two machines.
Fogel et a1 (1966) suggested drawing a single state diagram which expresses
the majority logic of an array of finite-state machines. Each state of the majority
logic machine is the composite of TEAM
a state
from each of the original machines.
LRN
285
Finite-state machines
0
0/1,1/0
0/0,1/0
110
W1
111
wi,i/n
WO
1/1
110
n/i
w1, 110
Figure 33.9. Three parent machines (top) are joined by a majority logic operator to
form another machine (bottom). The initial state of each machine is indicated by a short
arrow pointing to that state. Each state in the majority logic machine is a combination
of the states of the three parent machines with the output symbol being chosen as the
majority decision of two of the three parent machines. For example, the state BDF in
the majority logic machine is determined by examining the states B, D, and F in each
of the individual machines. For an input symbol of 0, all three states respond with a
0, therefore this symbol is chosen for the output to an input of 0 in state BDF. For an
input symbol of I , two of the three states respond with a 0, thus, this being the majority
decision, this symbol is chosen for the output to an input of 1 in state BDF. Note that
several states of the majority logic machine are isolated from the start state and could
never be expressed.
Thus the majority machine may have a number of states as great as the product
of the number of states in the original machines. Each transition of the majority
machine is described by that input symbol which caused the respective transition
in the original machines, and by that output symbol which results from the
majority element logic being applied to the output symbols from each of the
original machines (figure 33.9). If there are only two parents to recombine in
this manner, the majority logic machine reduces to the better of the two parents.
Zhou and Grefenstette ( 1 986) used recombination on finite-state automata
TEAM
LRN
applied to binary sequence induction
problems.
The finite-state automata were
286
Recombination
287
of the parse tree must have the correct number of subtrees below it, one for
each argument that the function requires.
Often in genetic programming, a simplification is made so that all functions
and terminals in the primitive language return the same data type. This is
referred to as the closure principle (Koza 1992). The effect is to reduce the
number of syntactic constraints on the programs so that the complexity of the
crossover operation is minimized.
The recursive structure of parse tree representations makes the definition of
crossover for tree representations that adhere to the above caveats surprisingly
simple. Cramer (1985) initially defined the now standard subtree crossover
for parse trees shown in figure 33.10. First, a random subtree is selected
and removed from one of the parents. Note that this leaves a hole in the
parent such that there exists a function that has a null value for one of its
parameters. Next, a random subtree is extracted from the second parent and
inserted at the point in the first parent where its subtree was removed. Now
the hole in the first parent is again filled. The process is completed by
inserting the subtree extracted from the first parent into the position in the
second parent where its subtree was removed. As long as only complete
subtrees are swapped between parents and the closure principle holds, this simple
crossover operation is guaranteed to produce syntactically valid offspring every
execution.
Typically, when evolving parse tree representations, a user-defined limit on
the maximum size of any tree in the population is provided. Subtree crossover
will often increase the size of a given parent such that, over a number of
generations, individuals in an unrestricted population may grow to swamp the
available computational resources. Given a user-defined restriction on subtree
size, expressed as a limit according to either the depth of a tree or the number of
nodes it contains, crossover must enforce this limit. When a crossover operation
is executed that creates one or more offspring that violate the size limitation, the
crossover operation is invalidated and the offspring are restored to their original
forms. What happens next is a matter of choice. Some systems will reject both
children and revert back to selecting two new parents. Other systems attempt
crossover repeatedly either until both offspring fall within the size limit or until
a specified number of attempts is reached. Given the nature of the crossover
operation, the likelihood of performing a valid crossover operation in a small
number of attempts, say five, is fairly good.
Koza ( 1992) popularized the use of subtree crossover for manipulating
parse tree representations in genetic programming. The subtree swapping
crossover of Koza (1992) shares much with the subtree crossover defined
by Cramer (1985) with a few minor differences. The foremost difference is
a bias introduced by Koza (1992) to limit the probability that a leaf node
is selected as the subtree from a parent during crossover. The reasoning
for this bias according to Koza (1992) is that, in most trees, the number
TEAM LRNto the number of nonleaf nodes.
of leaf nodes will be roughly equivalent
288
Recombination
Figure 33.10. An illustration of the crossover operator for parse trees. A subtree is
selected at random from each parent, extracted, and exchanged to create two offspring
trees.
Other representations
289
33.6
Other representations
Introduction
To make the following survey unambiguous we have to start with setting some
conventions on terminology. The term population will be used for a multiset
of individuals that undergoes selection and reproduction. This terminology
is maintained in genetic algorithms, evolutionary programming, and genetic
programming, but in evolution strategies all p individuals in a ( p , A ) or ( p + A )
strategy are called parents. We, however, use the term parents only for those
individuals that are selected to undergo recombination. In other words, parents
are those individuals that are actually used as inputs for a recombination
operator; the arity of a recombination operator is the number of parents it uses.
The next notion is that of a donor, being a parent that actually contributes to (at
least one of) the alleles of the child(ren) created by the recombination operator.
TEAM
This contribution can be for instance
theLRN
delivery of an allele, as in uniform
290
Recombination
Multiparent recombination
29 1
with probability
Using the value p = 0.8, the simplex GA performed better than the standard
GA on the DeJong functions. The authors remark that applying a modified
crossover on more than three parents 'is worth to try'.
The problem of placing actuators on space structures is addressed by Furuya
and Haftka ( 1993). The authors compare different crossovers: among others
they use uniform crossover with two as well as with three parents in a GA
using integer representation. Based on the experimental results they conclude
that the use of three parents did not improve the performance. This might be
related to another conclusion, indicating that for this problem mutation is an
efficient operator and crossover might not be important. Uniform crossover
with an arbitrary number of parents is also used by Aizawa ( 1 994) as part of a
special schema sampling procedure in a GA, but the multiparent feature is only
a side-effect and is not investigated.
A so-called triadic crossover is introduced and tested by Ph1 (1994) for a
multimodal spin-lattice problem. The triadic crossover is defined in terms of
two parents and one extra individual, chosen randomly from the population. The
operator creates one child; it takes the bits in positions where the first parent
and the third individual have identical bits from this parent and the rest of the
bits from the other parent. Clearly, the result is identical to the outcome of
a voting crossover on these three individuals as parents. Although the paper
is primarily concerned with different selection schemes, a comparison between
triadic, one-point, and uniform crossover is made, where triadic crossover turned
out to deliver the best results.
292
Recombination
=
x:
+ X I (x?
- x:
where the two parents xsi, sT (S,, T, E ( 1 , . . . , p } ) are redrawn for each i
anew and so is the contraction factor x l . The above definition applies to the
object variables as well as the strategy parameter part; that is, for the mutation
stepsizes (a)and the rotation angles (a).Observe that the multiparent character
of global recombination is the consequence of redrawing the parents xsi , x T for
each coordinate i . Therefore, probably more than two individuals contribute
to the offspring x , but their number is not defined in advance. It is clear
that investigations on the effects of different numbers of parents on algorithm
performance could not be performed in the traditional ES framework. The option
of using multiple parents can be turned on or off, that is, global recombination
can be used or not, but the arity of the recombination operator is not tunable.
Experimental studies on global versus two-parent recombination are possible,
but so far there are almost no experimental results available on this subject.
Schwefel (1995) notes that appreciable acceleration is obtained by changing,to
a bisexual from an asexual scheme (i.e., adding recombination using two parents
to the mutation-only algorithm), but only a slight further increase is obtained
when changing from bisexual to multisexual recombination (i.e., using global
recombination instead of the two-parent variant). Recall the remark on the name
p-sexual voting. The terms bisexual and multisexual are not appropriate either
for the same reason: individuals have no gender or sex, and recombination can
be applied to any combination of individuals.
Gene-pool recornbination (GPR) was introduced by Muhlenbein and Voigt
(1996) as a multiparent recombination mechanism for discrete domains. It is
defined as a generalization of two-parent recombination (TPR). Applying GPR
is preceded by selecting a gene pool consisting of would-be parents. Applying
GPR the two parent alleles of an offspring are randomly chosen for each locus
with replacement from the gene pool and the offspring allele is computed using
any of the standard recombination schemes for TPR. Theoretical analysis on
infinite populations shows that GPR is mathematically more tractable than TPR.
I f n stands for the number of variables (loci), then the evolution with proportional
selection and GPR is fully described by n equations, while TPR needs 2
equations for the genotypic frequencies. In practice GPR converges about 25%
fdster than TPR for Onemax. The authors conclude that GPR separates the
identification and the search of promising areas of the search space better;
besides it searches more reasonably than does TPR. Voigt and Muhlenhein
(1995) extend GPR to continuous domains by combining it with uniform fuzzy
two-parent recombination (UFTPR) from Voigt et a1 ( 1995). The resulting
uniform fuzzy gene-pool recombination (UFGPR) outperforms UFTPR on the
TEAM
LRN
heritability,
giving it a higher convergence
spherical function in terms of realized
Multiparent recombination
293
speed. The convergence of UFGPR is shown to be about 25% faster than that
of UFTPR.
A very particular mechanism is the linkage e\d\ting genetic operator
(LEGO) as defined by Smith and Fogarty (1996). The mechanism is designed
to detect and propagate blocks of corresponding genes of potentially varying
length during the evolution. Punctuation marks in the chromosomes denote
the beginning and the end of each block and more chromosomes with the
appropriately positioned punctuation marks are considered as donors of a whole
block during the creation of a child. Although the multiparent feature is only a
side-effect, LEGO is a mechanism where more than two parents can contribute
to an offspring.
TEAM LRN
294
Recombination
-
:= .T
{ I , . . . ,r } ;
I/
i, := min{k I k 2 i , , x i @
{XI,
. . . , x,,}}
j E { I , . .. ,r } .
Observe that, because of the term k 2 i, above, a marker can remain at the
same position after an UPDATE, and will only be shifted if the allele standing
at that position is included in the child. This guarantees that each offspring will
be a permutation.
Depending on the mechanism of choosing a parent (and thereby an allele)
there are three different versions of scanning. The choice can be deterministic,
choosing a parent containing the allele with the highest number of occurrences
and breaking ties randomly (oc.c.itrrerzce-6asedsc*annirig). Alternatively it can
be random, either unbiased, following a uniform distribution thus giving each
TEAM
LRN
parent an equal chance to deliver its
allele
(imfiwrn scanning), or biased bj the
Multiparent recombination
295
fitness of the parents, where the chance of being chosen is fitness proportional
(fitness-based scanning). Uniform scanning for r = 2 is the same as uniform
crossover, although creating only one child, and it also coincides with discrete
recombination in evolution strategies. The occurrence-based version is very
much like the voting or majority mating mechanism discussed before. but
without the threshold v or with v = [ m / 2 J respectively. The effect of the
number of parents in scanning crossover has been studied in several papers. An
overview of these studies is given in the next subsection.
Diagorzal crossover has been introduced as a generalization of one-point
crossover in GAS (Eiben et a1 1994). In its original form diagonal crossover
creates r children from r parents by selecting r - 1 crossover points in the parents
and composing the children by taking the resulting r chromosome segments from
the parents along the diagonals. Later on, a one-child version was introduced
(van Kemenade et a1 1995). Figure 33.1 1 illustrates both variants. It is easy to
see that for r = 2 diagonal crossover coincides with one-point crossover, and in
some sense it also generalizes traditional two-parent n-point crossover. To be
precise, if we define ( r ,s) segmentation crossover as working on r parents with
J crossover points, diagonal crossover becomes its ( r , r - 1) version, its (2. n )
variant coincides with n-point crossover, and one-point crossover is an instance
of both schemes for ( r , s ) = (2, 1 ) as parameters. The effect of operator arity
for diagonal crossovers will be also discussed in the next subsection.
A recombination mechanism with tunable arity in ES is proposed by
Schwefel and Rudolph (1995). The ( p , K , A, p ) ES provides the possibility
of freely adjusting the number of parents (called ancestors by the authors).
The parameter p stands for the number of parents and global recombination is
redefined for any given set { X I , . . . , x P ) of parents as
=
;j/p)cP
k=l
x,k
296
Recombination
Figure 33.11. Diagonal crossover (top) and its one-child version (bottom) for three
parents.
MultiDarent recombination
297
<x:>"',
X y '
( x y
...
+ +
In recent years quite a few papers have studied the effect of operator arity on EA
performance, some even in combination with varying selective pressure. Here
we give a brief summary of these results, sorted by articles.
The performance of scanning crossover for different numbers of parents is
studied by Eiben et a1 (1994) in a generational GA with proportional selection.
Bit-coded GAS for function optimization (DeJong functions FI-4 and a function
from Michalewicz) as well as order-based GAS for graph coloring and the TSP
are tested with different mechanisms to CHOOSE. In the bit-coded case more
parents perform better than two; for the TSP and graph coloring two parents
are advisable. Comparing different biases in choosing the child allele, on four
out of the five numerical problems fitness-based scanning outperforms the other
two and occurrence-based scanningTEAM
is theLRN
worst operator.
298
Recombinati(E
Multiparent recombination
299
Figure 33.12. Illustration of the effect of the number of parents (horizontal axis) on the
error at termination (vertical axis) on NK landscapes with NNI, N = 100, K = 1 (top),
K = 25 (bottom).
recombination. It seems that if and when crossover is useful, that is, on mildly
epistatic problems, then multiparent crossover can be more useful than the twoparent variants.
The results of an extensive study of diagonal crossover for numerical
optimization in GAS are reported by Eiben and van Kemenade (1997). Diagonal
TEAM LRN
crossover is compared to its one-offspring
version and n-point crossover on a test
300
Recombination
suite consisting of eight functions, monitoring the speed, that is, the total number
of evaluations, the accuracy, that is, the median of the best objective function
value found (all functions have an optimum of zero), and the success rate, th;it
is, the percentage of runs where the global optimum is found. In most cases a n
increase of performance can be achieved by increasing the disruptivity of the
crossover operator (using higher values of n for n-point crossover), and even
more improvement is achieved if the disruptivity of the crossover operator and
the number of parents is increased (using more parents for diagonal crossover).
This study gives a strong indication that for diagonal crossover an advantageous
multiparent effect does exist, that is, (i) using this operator with more than tmo
parents increases GA performance and (ii) this improvement is not only the
consequence of the increased number of crossover points.
A recent investigation of Eiben and Back (1097) addresses the working of
multiparent recombination operators in continuous search spaces, in particulz
within ESs. This study compares p / 2 intermediate recombination, pary discrete
recombination, which is identical to uniform scanning crossover, and diagonal
crossover with one child. Experiments are performed on unimodal landscapes
(sphere model and Schwefels double sum), multimodal functions with regularly
arranged optima and a superimposed unimodal topology (Ackley, Griewangk,
and Rastrigin functions) and on the Fletcher-Powell and the Langermarin
functions that have an irregular, random arrangement of local optima. On
the Fletcher-Powell function multiparent recombination does not increase
evolutionary algorithm (EA) performance; besides for the unimodal double sum
increasing operator arity decreases performance. Other conclusions seem to
depend on the operator in question; the greatest consequent improvement on
raising the number of parents is obtained for diagonal crossover.
33.7.6 Conclusions
The idea of applying more than two parents for recombination in an evolutionary
problem solver occurred as early as the 1960s (Bremermann et a1 1966). Several
authors have designed and applied recombination operators with higher arities
for a specific task, or used an existing operator with an arity higher than two
(Kaufman 1967, Muhlenbein 1989, Bersini and Seront 1992, Furuya and Haftka
1993, Aizawa 1994, Pal 1994). Nevertheless, investigations explicitly devoted
to the effect of operator arity on EA performance are still scarce; the study of the
phenomenon of multiparent recombination has just begun. What would such a
study mean? Similarly to the question of whether binary reproduction operators
(crossover with two parents) have advantages over unary ones (using mutation
only), it can be investigated whether or not using more than two parents is
advantageous. In the case of operators with tunable arity this question can be
refined and the relationship between operator arity and algorithm performance
can be studied. It is, of course, questionable whether multiparent recombination
TEAM LRN showing one behavioral pattern.
can be considered as one single phenomenon
Multiparent recombination
30 1
The survey presented here discloses that there are (at least) three different types
of multiparent mechanism with tunable arity:
(0
A priori it cannot be expected that these different schemes show the same
response to raising operator arities. There are also experimental results
supporting differentiation among various multiparent mechanisms. For instance,
there seems to be no clear relationship between the number of parents and
the performance of uniform scanning crossover, while the opposite is true for
diagonal crossover (Eiben and Schippers 1996).
Another aspect multiparent studies have to take into consideration is the
expected different behavior on different types of fitness landscape. As no
single technique would work on every problem, multiparent mechanisms will
have their limitations too. Some studies indicate that on irregular landscapes.
such as NK landscapes with relatively high K values (Eiben and Schippers
1996), or the Fletcher-Powell function (Eiben and Back 1997). they do not
work. On the other hand, on the same Fletcher-Powell function Eiben and van
Kemenade ( I 997) observed an advantage of increasing the number of parents for
diagonal crossover in a GA framework using bit coding of variables, although
they also found indications that this can be an artifact, caused simply by the
increased disruptiveness of the operator for higher arities. Investigations on
multiparent effects related to fitness landscape characteristics smoothly fit into
the tradition of studying the (dis)advantages of two-parent crossovers under
different circumstances (Schaffer and Eshelman 1991, Eshelman and Schaffer
1993, Spears 1993, Hordijk and Manderick 1995).
Let us also touch on the issue of practical difficulties when using multiparent
recombination operators. Introducing operator ari ty as a new parameter implies
an obligation of setting its value. Since so far there are no reliable heuristics for
setting this parameter, finding good values may require numerous tests, prior
to 'real' application of the EA. A solution may be based on previous work on
adapting (Davis 1989) or self-adapting (Spears 1995) the frequency of applying
different operators. Alternatively, a number of competing subpopulations could
be used in the spirit of Schlierkamp-Voosen and Miihlenbein (1996). According
to the latter approach each different arity is used within one subpopulation
and subpopulations with greater progress, that is, with more powerful operators.
LRN
become larger. A first assessment ofTEAM
this technique
can be found in an article by
302
Recombinatior!
Eiben er al ( 1998a). Another recent result indicates the advantage of using more
parents in the context of constraint satisfaction problems (Eiben et a1 1998b).
Concluding this survey we can note the following. Even though there are
no biological analogies of recombination mechanisms where more than two
parent genotypes are mixed in one single recombination act, formally there is
no necessity to restrict the arity of reproduction mechanisms to one (mutation)
or two (crossover) in computer sirnulations. Studying the phenomenon of
multiparent recombination has just begun, but there is already substantial
evidence that applying more than two parents can increase the performancc
of EAs. Considering multiparent recombination mechanisms is thus a sound
design heuristic for practitioners and a challenge for theoretical analysis.
References
Ackley D H 1987a A Connectionist Machine for Genetic Hillclimbing (Boston, MA:
Kluwer)
__ 1987b An empirical study of bit vector function optimization Genetic Algorithms and
Simulated Anneulirtg ed L Davis (San Mateo, CA: Morgan Kaufmann) pp 170-2 I5
Aizawa A N 1994 Evolving SSE: a stochastic schemata exploiter Proc. 1st IEEE Con$ on
Evolutionun Computation (Orlundo, FL, 1990) (Piscataway, NJ: IEEE) pp 5 2 5 4
Altenberg L 1995 The schema theorem and Prices theorem Foundations c f l Genetic
Algorithms 3 ed L Whitley and M Vose (San Mateo, CA: Morgan Kaufmann)
Back T 1996 Evolutionun Algorithms in Theory und Pruc e (New York: Oxfotd
University Press )
Back T, Rudolph G and Schwefel H-P 1993 Evolutionary programming and evolution
strategies: similarities and differences Proc. 2nd Ann. Con$ on Evolutionuiy
Progrumnzing (Sun Diego, CA) ed D B Fogel and W Atmar (La Jolla, CA:
Evolutionary Programming Society) pp I 1-22
Back T and Schutz M 1995 Evolution strategies for mixed-integer optimization of optical
multilayer systems Proc. 4th Anti. Conf: on Evolutionury Prugrumming (Sun Diego,
CA, March 1995) ed J R McDonnell, R G Reynolds and D B Fogel (Cambridge,
MA: MIT Press) pp 33-51
Bersini H and Seront G 1992 In search of a good evolution-optimization crossover
Purullel Problem Sohing frum Nuture, 2 (Proc. 2nd lnt. Conf on Purullel Problem
Soliling jrom Nuture, Brussels, 1992) ed R Manner and B Manderick (Amsterdam:
Elsevier-North-Holland) pp 479-88
Beyer H-G 1995 Toward a theory of evolution strategies: on the benefits of sex-the
( p / p ,A ) theory Etvlutionury Coniput. 3 81-1 1 1
-I996 Basic principles for a unified EA-theory E\dutionury Algorithms and their
Applicutions Workshop (Dugstuhl, 1996)
Birgmeier M I996 Evolutionary programming for the optimization of trellis-coded
modulation schemes Pruc. 5th Ann. Con$ on Evolutionq Prugramming ed L J
Fogel, P J Angeline and T Back (Cambridge, M4: MIT Press)
Blanton J and Wainwright R 1993 Multiple vehicle routing with time and capacity
constraints using genetic algorithms Proc. 5th Int. Cmj; on Genetic Algorithms
(Urburzci-ChcinipLiigri, IL, July 1093) ed S Forrest (San Mateo, CA: Morgan
TEAM LRN
Kaufmann) pp 452-9
References
303
304
Recombinatioil
References
305
306
Recombination
References
307
Syswerda G 1989 Uniform crossover in genetic algorithms Proc. 3rd Int. Con$ on Genetic
Algorithms (Fairjkx, VA, June 1989) ed J D Schaffer (San Mateo, CA: Morgan
Kaufmann) pp 2-9
-1
99 1 Schedule optimization using genetic algorithms Handbook of Genetic
Algorithms ed L Davis (New York: Van Nostrand Reinhold) pp 3 3 2 4 9
Thierens D and Goldberg D E 1993 Mixing in genetic algorithms Proc. 5th Int. Con$ on
Genetic Algorithms (Urbana-Chumpaign, IL, July 1993) ed S Forrest (San Mateo,
CA: Morgan Kaufmann) pp 3 8 4 5
van Kemenade C, Kok J and Eiben A 1995 Raising GA performance by simultaneous
tuning of selective pressure and recombination disruptiveness Proc. 2nd IEEE Con$
on Evolutionary Computation (Perth, 1995) (Piscataway, NJ: IEEE) pp 346-5 1
Voigt H-M and Muhlenbein H 1995 Gene pool recombination and utilization of
covariances for the breeder genetic algorithm Proc. 2nd IEEE Int. Con$ on
Evolutionary Computation (Perth, 1995) (Piscataway, NJ: IEEE) pp 172-7
Voigt H-M, Muhlenbein H and CvetkoviC D 1995 Fuzzy recombination for the breeder
genetic algorithm Proc. 6th Int. Con$ on Genetic Algorithms (Pittsburgh, PA, 1995)
ed L J Eshelman (San Mateo, CA: Morgan Kaufmann) pp 1 0 4 - 1 1
Whitley D, Starkweather T and Shaner D 1991 Traveling salesman and sequence
scheduling: quality solutions using genetic edge recombination Handbook of
Genetic Algorithms ed L Davis (New York: Van Nostrand Reinhold) pp 350-72
Whitley D and Yoo N-W 1995 Modeling simple genetic algorithms for permutation
problems Foundations of Genetic Algorithms 3 ed D Whitley and M Vose (San
Mateo, CA: Morgan Kaufmann) pp 163-84
Wright A H 1994 Genetic algorithms for real parameter optimization Foundations of
Genetic Algorithms ed G Rawlins (San Mateo, CA: Morgan Kaufmann) pp 205-1 8
Wu A S and Lindsay R K 1995 Empirical studies of the genetic algorithm with noncoding
segments Evolutionary Comput. 3 121 4 8
Zhou H and Grefenstette J J 1986 Induction of finite automata by genetic algorithms Proc.
1986 IEEE Int. Con5 on Systems, Man, and Cybernetics (Atlanta, GA) pp 170-4
TEAM LRN
Other operators
Russell W Anderson (34.1), David B Fogel (34.2) and
Martin Schiitz (34.3)
34.1
Russell W Andersont
34.1.1 Interactions between learning and evolution
308
TEAM LRN
309
310
Other operators
31 1
Combinations of alleles
Figure 34.1. Schematic representation of the fitness landscape in the model of Hinton
and Nowlan. A two-dimensional representation of genome space in the problem
considered by Hinton and Nowlan (1987). The horizontal axis represents all possible
gene combinations, and the vertical axis represents relative fitness. Without learning, only
one combination of alleles correctly completes the network; hence only one genotype has
higher fitness, and no gradient exists. The presence of plastic alleles radically alters this
fitness landscape. Assume a correct mutation occurs in one of the 20 genes. The advent
of a new correct gene only partially solves the problem. Learning allows individuals
close (in Hamming space) to complete the solution. Thus, these individuals will be
slightly more fit than individuals with no correct genes. Useful genes will thereby be
increased in subsequent generations. Over time, a large number of correct genes will
accumulate in the gene pool, leading to a completely genetically determined structure.
1991, 1994, Whitley and Gruau 1993, Whitley et a1 1994, Balakrishnan and
Honavar 1995, Turney 1995, 1996, Turney et a1 1996). Considering the rather
specific assumptions of their model, it is useful to contemplate which aspects
of their results are general properties. Among the issues raised by this and
subsequent studies are the degree of biological realism, the nature of the fitness
landscape, the computational cost of learning, and the role of learning in static
fitness landscapes.
First, the models assumption of plastic alleles that can mutate into
permanent alleles seems biologically spurious. However, the Baldwin effect
can be manifested in the evolution of a biological structure regardless of
the genetic basis of that structure or the mechanisms underlying the learning
process (Anderson 1995a). The Baldwin effect is simply a consequence of
individual learning on genetic evolution. Subsequent studies have demonstrated
the Baldwin effect using a variety of learning algorithms. Turney (1995, 1996)
has observed a Baldwin effect in a class of hybrid algorithms, combining a
genetic algorithm (GENESIS) and an inductive learning algorithm, where the
Baldwin effect was manifested in shifting biases in the inductive learner. French
and Messinger (1994) investigated the Baldwin effect under various forms of
TEAM
LRN observed the Baldwin effect in a
phenotypic plasticity. Cecconi et al
(1995)
312
Other operators
313
1989, Scheiner 1993, Via 1993) as well as for sexual versus asexual reproduction
(Maynard-Smith 1978).
34.1.3 Quantitative genetics models
In order to make some of these issues more explicit, it is useful to study
the Baldwin effect under the general assumptions of quantitative genetics.
A quantitative genetics methodology for modeling the effects of learning on
evolution was developed by Anderson (1995a), and the primary results of this
analysis are reviewed in this section. The limitations of this theoretical approach
are well known. For example, quantitative genetics assumes infinite population
sizes. Also, complete analysis is often limited to a single quantitative character.
Nevertheless, such analyses can provide a baseline intuition regarding the effects
of learning and evolution.
All essential elements of an evolutionary process subject to the Baldwin
effect are readily incorporated into a quantitative genetics model. These
elements include (i) a function for the generation of new genotypes through
mutation and/or recombination, (ii) a mapping from genotype to phenotype, (iii)
a model of the effects of learning on phenotype, and (iv) a selection function.
In this section, this methodology is demonstrated for a simple, first-order model,
where only the phenomenological effects of learning on selection are considered.
More advanced models are discussed, which incorporate a model of the learning
process, along with its associated costs and benefits. These analyses illustrate
several underappreciated points: (i) learning generally slows genetic change, (ii)
learning offers no long-term selective advantage in fixed environments, and (iii)
the effects of learning are somewhat independent of the mechanisms underlying
the learning process.
Learning as a phenotypic variance. For a first-order model, consider an
individual whose genotype is a real-valued quantitative character subject to
normal (Gaussian) selection:
(34.1)
where w,(g) represents selection as a function of genotype, ge represents the
optimal genotype, and V , ( t ) is variance of selection as a function of time. A
direct mapping from genotype to phenotype is implicitly assumed.
What effect does learning have on this selection function? Learning allows
an individual to modify its phenotype in response to its environment. Consider
an individual whose genotype ( g i ) is a given distance (Igi - gel) from the
environmental optimum ( g e ) . Regardless of the mechanisms underlying the
learning process, the net effect of learning is to reduce the fitness penalty
associated with this genetic distance. Because of its ability to learn, an
LRN
individual with genotype gi has aTEAM
probability
of modifying its phenotype to
3 14
Other operators
The population mean and variance after selection (m*,V,*) can now be
expressed in the form of dynamic equations:
(34.5)
(34.6)
Lastly, mutations are introduced in the production of the next generation of
trials. To model this process, assume a Gaussian mutation function with mean
zero and variance VM. A convolution of the population distribution with the
mutation distribution has the effect of increasing the population variance:
where
(34.8)
Hence, in a fixed environment the population mean m ( t ) will converge on
the optimal genotype (Bulmer 1985), while a mutation-selection equilibrium
variance occurs at
vp (v; 4v, v y 2
v**
=
(34.9)
Peq
2
Inspection of equations ( 3 4 3 , (34.6), and (34.8) illustrates two important points.
TEAM
First, learning slows the convergence
of LRN
both m*(t) and V;(t). Second, once
315
convergence in the mean is complete, the utility of learning is lost, and learning
only reduces fitness.
In a more elaborate version of this model, called the critical learning period
model (Anderson 1995a), a second gene is introduced to regulate the fraction of
an individuals life span devoted to learning (duration of the learning period).
Specification of a critical learning period implicitly assigns cost associated with
learning (the percent of life span not devoted to reproduction). Individuals are
then selected for the optimal combination of genotype and learning investment.
It is easily demonstrated that under these assumptions, learning ability is selected
out of a population subject to fixed selection.
Constant-velocity environments. Next, consider a simple case of changing
selection-a constantly moving optimum, ge(t) = 6 t , where 6 is defined as
the environmental velocity. Let the difference between the population mean
and the environmental optimum be defined as @ = m ( t ) - g , ( t ) . The dynamic
equation for @ is
(34.10)
At equilibrium, @ * ( t )= @ ( t ) ,hence
@eq
vp
~
+ v; 6
(34.11 )
VP
316
Other operators
Baldwins essential insight was that if an organism has the ability to learn,
it can exploit genes that only partially determine a structure-increasing the
frequencies of useful genes in subsequent generations. The Baldwin effect has
also been demonstrated to be operative in hybrid evolutionary algorithms. These
TEAM
empirical investigations can be used
to LRN
quantify the benefits of incorporating
Knowledge-augmented operators
317
= 1.224 f (x)/II
318
Other operators
about the function being searched i n order to provide the greatest expected rate
of convergence. In this particular case, however, knowledge that the function is
a quadratic surface indicates the use of search algorithms that can take greater
advantage of the available gradient information (e.g. Newton-Gauss).
There are other instances where incorporating domain-specific knowledge
into a search operator can improve the performance of an evolutionary algorithm.
In the traveling salesman problem, under the ob-jective function of minimizing
the Euclidean distance of the circuit of cities, and a representation of simply m
ordered listing of cities to be visited, Fogel (1988) offered a mutation operator
which selected a city at random and placed it in the list at another randomly
chosen position. This operator was not based on any knowledge about the
nature of the problem. In contrast, Fogel (1993) offered an operator that instead
inverted a segment of the listing (i.e. like a 2-opt of Lin and Kernighan (1976)).
The inversion operator in the traveling salesman problem is a knowledgeaugmented operator because it was devised to take advantage of the Euclidean
geometry present in the problem. In the case of a traveling salesmans tour, if
the tour crosses over itself it is always possible to improve the tour by undoing
the crossover (i.e. the diagonals of a quadrangle are always longer in sum than
any two opposite sides). When the two cities just before and after the crossing
point are selected and the listing of cities in between reversed, the crossing is
removed and the tour is improved. Note that this use of inversion is appropriate
in light of the traveling salesman problem, and no broader generality of its
effectiveness as an operator is suggested, or can be defended.
Domain knowledge can also be applied in the use of recombination. For
example, again when considering the traveling salesman problem, Grefenstette et
N I (1985) suggested a heuristic crossover operator that could perform a degree
of local search. The operator constructed an offspring from two parents by
(i) picking a random city as the starting point, (ii) comparing the two edges
leaving the starting cities in the parents and choosing the shorter edge, then (iii)
continuing to extend the partial tour by choosing the shorter of the two edges
in the parents which extend the tour. If a cycle were introduced, a random
edge would be selected. Grefenstette et a1 (1985) noted that offspring were
on average about 10% better than the better parent when implementing this
operator.
In many real-world applications, the physics governing the problem suggests
settings for search parameters. For example, in the problem of dockmg
small molecules into protein binding sites, the intermolecular potential can be
precalculated on a grid. Gehlhaar et a1 (1995) used a grid of 0.2 A, with each
grid point containing the summed interaction energy between an atom at that
point and all protein atoms within 6 A. This suggests that under Gaussian
perturbations following an evolutionary programming or evolution strategy
approach, a standard deviation of several hgstroms would be inappropriate
(i.e. too large).
TEAM
Whenever evolutionary algorithms
areLRN
applied to specific problems with the
319
34.3.2 Basic motivations for the use of gene duplication and deletion
From these first attempts concerning variable-length genotypes until now many
researchers have made use of gene duplication and deletion. Four different
motivations may be classified.
(i) Engineering applications. Many difficult optimization tasks arise from
engineering applications in which variable-dimensional mixed-integer problems
have to be solved. Often these problems are of dynamic nature: the optimum
is time dependent. Additionally, in order to obtain a reasonable model of
the system under consideration, a large number of constraints has to be
respected during the optimization. Solving the given task frequently assumes
the integration of expert (engineer) knowledge into the problem solving strategy:
into particular genetic operators in the case of EAs. Many such constrained,
variable-dimensional, mixed-integer, time-varying engineering problems and
their solutions can be found in the handbook by Davis (1991) and in the
TEAM
LRN
proceedings of several conferences,
such
as the International Conference on
320
Other operators
32 I
(iv) ArtiJcial intelligence. Another important field in which variabledimensional techniques have also been used is the domain of artificial
intelligence (AI), especially machine learning (ML) and artijicial life (AL).
Whereas in the field of ML (subordinated fields are, for example, genetic
programming, classifier systems, and artificial neural networks) solving a
possibly variable-dimensional optimization problem (depending on the actual
subordinated field in mind) is one main objective, this aim plays a minor
role in the AL field. AL research concentrates on computer simulations of
simple hypothetical life forms and selforganizing properties emerging from
local interactions within a large number of basic agents (life forms). A second
objective of AL is the question of how to make the agents' behavior adaptive,
thus often leading to agents equipped with internal rules or strategies determining
their behavior. In order to learn/evolve such rule sets, learning strategies such
as EAs are used. Since the number of necessary rules is not known a priori,
a variable-dimensional problem instance arises. Despite the rule learning task,
the modeling of the simple life forms itself makes use of variable-dimensional
genotypes.
V i E { l ,..., m }
V j E { l , ..., I }
hj(z)=O
2 = ( x i , .. . ,Xnr) E D C X
nE
g;(z)LO
f,g ; , hj : X -+ R.
322
Other operators
323
formalized as follows:
dup : I + I , with dup(a) = dup(x1, . . . , s,.. . . , x,,, s)
s) = a.
= ( X I , . . . ,X I , XI, . . . ,
Analogously to deletion a duplication probability Pdup E (0, 1 ) is used and the
index i is usually uniformly chosen. Concerning the policy for introducing the
new gene xf several policies may be distinguished, such as:
0
Duplication. The gene x,! is a duplicate of x,, such that a has the form
a = ( X I , . . . , X I , x;, . . . , x,, s).
Related. The initialization of the new gene xf is context dependent: xfis
generated with help the of the actual values of x , and .rr+l.
Addition. x, is initialized at random.
324
Other operators
34.3.5 Solutions
The e i d u t i o n program approach of Michalewicz ( 1992), i.e. combining
the concept of evolutionary computation with problem-specific chromosome
structures and genetic operators, may be seen as one main concept used to
overcome the problems mentioned above. Although this concept is useful in
practice, it prevents the conception of a more general and formalized view
of variable-length EAs because there no longer exists the EA using the
representation and the set of operators, Instead, for each application problzm
a specialized EA exists. According to Lohmann (1992) and Kost (1993),
for example. the formulation of operators such as gene duplication and
deletion, used in their framework of structural evolution, is strongly application
dependent, thus inhibiting a more formal, general concept of these operators.
Davidor ( I 99 la, b) expressed the need for revised and new genetic operators
for his variable-length robot trajectory optimization problem. In contrast to
the evolution program approach, Schutz ( 1994) formulated an applicationindependent, variable-dimensional mixed-integer evolution strategy (ES), thus
following the course of constructing a more general sort of ES. This offered
Schutz the possibility to be more formal than other researchers. Unfortunately,
this approach is restricted to a class of problems which can easily be mapped
onto the mixed-integer representation he used.
Because most work concerning variable-length genotypes uses the evolution
program approach, a formal analysis of gene duplication and deletion is rarely
found in the literature and is therefore omitted here. As a consequence,
theoretical knowledge about the behavior of gene duplication and deletion is
nearly unknown. Harvey ( 19931, for example, points out that gene-duplication,
followed by mutation of one of the copies, is potentially a powerful method for
evolutionary progress. Most statements concerning nonstandard operators such
as duplication and deletion have the same quality as Harveys: they are far from
being provable.
Because of the lack of theoretical knowledge we proceed by discussing
some solutions used to circumvent the problems which arise when introducing
variable-length genotypes. In the first place, we question how other researchers
have solved the problem of noncomparable loci, i.e. the problem of respecting
the semantics of loci. Mostly this gene assignment problem is solved by
explicitly marking semantical entities on the genotype. The form of the tagging
varies from application to application and is carried out with the help of different
representations.
0
325
Harp and Samad (1991) implemented the tagging with the help of a special
and more complex data structure representing the structure and actual
weights of any feedforward net consisting of a variable number of hidden
layers and a variable number of units.
Goldberg et a1 ( I 989, 1990) extended the usual string representation of
GAS by using a list of ordered pairs, with the first component of each tuple
representing the position in the string and the second one denoting the
actual bit value. Using genotypes of fixed length a variable dimension
in the resulting messy GA was achieved by allowing strings not to
contain full gene complement (underspecification) and redundant or even
contradictionary genes (overspecification).
Koza ( 1992, 1994) used rooted point-labeled trees with ordered branches
(LISP expressions), thus having a genotype representing semantics very
well.
Lohmann ( 1992) circumvented the assignment problem using so-called
structural evolution. The basic idea of structural evolution is the separation
of structural and nonstructural parameters, thus leading to a two-level ES:
a multipopulation ES using isolation. While on the level of each population
a parameter optimization, concerning a fixed structure, is carried out, on
the population level several isolated structures compete with each other. In
this way Lohmann was able to handle structural optimization problems with
variable dimension: the dimension of the structural parameter space does
not have to be constant. Since each ES itself worked on a fixed number
of nonstructural parameters (here a vector of reals) no problem occurred on
this level. On the structural level (population level) special genetic operators
and a special selection criterion were formulated. The criticism concerning
structural evolution definitively lies in the basic assumption that structural
and nonstructural parameters can always be separated. Surely, many mixedinteger variable-dimensional problems are not separable. Secondly, on the
structural level the well-known semantical problem exists, but was not discussed.
Schutz (1 994) totally omitted a discussion concerning semantical problems
arising from variable-length genotypes.
If the genotype is sufficiently prepared, problems (especially) concerning
recombination disappear, because the genetic operators may directly use the
tagging in order to construct interpretable individuals. Another important idea
when designing recombination operators for variable-length genotypes is pointed
out by Davidor (1991a). He suggests a matching of parameters according to
their genotypic character instead of to their genotypic position. Essentially, this
leads to a matching on the phenotypic, instead of the genotypic level. Generally,
Davidor points out:
326
Other operators
occurs between sites that control the same, or at least the most similar,
function in the phenotypic space.
One may remark that many ideas concerning the use of gene duplication ;ind
deletion exist. Unfortunately, most thoughts have been extremely application
oriented, that is, not formulated generally enough. Probably the construction of
a formal frame will be very complicated in the fBce of the diversity of problems
and solutions.
References
Ackley D and Littman M 1991 Interactions between learning and evolution ArriJlciu[
Lfe I1 (Suntci Fe, N M , Fehruuq 1990) ed C Langton, C Taylor, D Farmer and S
Rasmussen (Redwood City, CA: Addison-Wesley) pp 487-509
-1994
A case for Lamarckian evolution Artijicicil Lfe 111 ed C Langton (Redwood
City, CA: Addison-Wesley) pp 3-10
Anderson R W 1995a Learning and evolution: a quantitative genetics approach J. Theor.
Biol. 175 89-101
-1
99% Genetic mechanisms underlying the Baldwin effect are evident in natural
antibodies Proc. 4th A m . COY$on Esolutionrzn' Progrutnming (Sun Diego, CA,
March 1995) ed J R McDonnell, R G Reynolds and D B Fogel (Cambridge, MA:
MIT Press) pp 547-63
-1996a How adaptive antibodies facilitate the evolution of natural antibodies
Immunol. Cell Biology 74 286-9 I
-1996b Random-walk learning: a neurobiological correlate to trial-and-error Prog.
Neurul Nemvrk.5 at press
Back T 1996 E\dutioticin Algorithms in Theon cirid Prac-tic-e (New York: Oxford
University Press)
Balakrishnan K and V Honavar I995 Ewlutionury Desigri of Neurul Architectiires:
(1 Prelimitiury k o n o t n y und Guide to Literature Artificial Intelligence Research
Group, Department of Computer Science, Iowa State University, Technical Report
CS TR 95-01
Baldwin J M 1896 A new factor in evolution Am. Nuturulist 30 44 1-5 I
Belew R K 1989 When both individuals and populations search: adding simple learning
to the genetic algorithm Proc. 3rd Int. Con$ on Genetic Algorithms (Fuitjiix, VA,
Jirne 1989) ed J D Schaffer (San Mateo, CA: Morgan Kaufmann) pp 3 4 4 1
-1
990 Evolution, learning and culture: computational metaphors for adaptive search
Cot?lples Syst. 4 1 1-49
Bremermann H J and Anderson R W 1991 How the brain adjusts synapses-maybe
Aittomated Reasoning: E s s a y in Horior of Woody Bledsoe ed R S Boyer (New
TEAM LRN
York: Kluwer) pp 119-47
References
327
328
Other operators
Hightower R, Forrest S and Perelson A 1996 The Baldwin effect in the immune system:
learning by somatic hypermutation Aduptirte Itidiriducils in E r d r i n g Populcition.\:
Models Lind Algorithms ed R K Belew and M Mitchell (Reading, MA: AddisonWesley) at press
Hiriton G E and Nowlan S J 1987 How learning can guide evolution Comp1e.x Syst. 1
495-502
Holland J H 1975 A~lciptcitioti in Niitiird r i d Artlficid S y s t e m (Ann Arbor, M I :
University of Michigan Press)
Kost B I993 Strict-tiircil Design iiti Er~)lutionStrcrtegic>s Internal Report, Department of
Bionics and Evolution Technique, Technical University of Berlin
Koza J R 1992 Genetic Progrcimniing (Cambridge, MA: MIT Press)
-1
994 Genetic Progrcltnmittg lI (Cambridge, MA: MIT Press)
Lin S and Kernighan B W 1976 An effective heuristic for the traveling salesman problem
Opercit. Res. 21 498-5 16
Lohmann R 1992 Structure evolution and incomplete induction Pcircillel Problem Solrilzg
from Nature. 2 (Proc. 2nd lnt. Cot$ on Peirullel Probletit Soh9ing from Nciticm,
Briissel\, 1992) ed R Manner and B Manderick (Amsterdam: Elsevier) pp 1 7 5 4 5
Manner R and Manderick B (ed5) 1992 Parctllel Problem Sohing jrom Nritiire, 2 (Proc..
2nd Int. Cot$ on Prirrillel Problem Sohirig from Nutitre, Br14.\,\d.\, 1992) ed R
Manner and B Manderick (Amsterdam: Elsevier)
Maynard Smith J 1978 The E\diction of Sex (Cambridge: Cambridge Univer\ity Pres,)
_- 1987 When learning guides evolution Ncitirre 329 76 1-2
Michalewicz Z I992 Genetic Algorithms
Dutci S t r i ~ t u r e s = E\dittioii Progrcirns
(Berlin: Springer)
Milstein C 1990 The Croonian lecture 1989 Antibodies: a paradigm for the biology of
molecular recognition Proc. R. Soc. B 239 1-16
Mitchell M and Belew R K 1995 Preface to G E Hinion and S J Nowlan How learning
can guide evolution A@ t i rte Int li \licluci Is it z E\ vol ring Pop it lu tions : Moclels ( I tzd
Algorithtns ed R K Belew and M Mitchell (Reading, MA: Addison-Wedey)
Morgan C L 1896 On modification and variation Science 4 733-40
Osborn H F 1896 Ontogenic and phylogenic variation Science 4 786-9
Nolfi S. Elman J L and Parisi D 1094 Learning and evolution in neural network5 Adcipt/\-e
BeCiurior 3 5-28
Paechter B, Cumming A, Norman M and Luchian H 1995 Extensions to a memetic
timetabling sy\tem Proc. 1st Int. Conf: oiz the t'rcictice cirzd Tlzeon cf Auto,ncired
fimetublirzg (ICPTAT 95) (Edinburgh, 1995)
Parisi D, Nolfi S and Cecconi F 1991 Learning, behavior, and evolution finrvird c1 Prcictrce
of Aittonotnoits S y s t e m (Proc. 1st Eur. Conj: on Artijicicil L f e (Puris, 1991)) ed F
J Varela and P Bourgine (Cambridge, MA: MIT Press)
Rechenberg I 1973 Ei~olittioiis.\trciteRie: Optinzieritrig Technisctier Systerne n d i
Principien der Biologi.\c*hrtz Erdirtion (Stuttgart: Fromman-HolLboog)
Saravanan N, Fogel D B and Nelson K M 1995 A comparison of methods for selfadaptation in evolutionary algorithms BioSystern\ 36 157-66
Scheiner S M 1993 Genetic\ and evolution of phenotypic plasticity Ann. Re\-. c o l .
Sv\temcit. 24 35-68
S c h Ut7 M I 994 Eirie Er v lu tions J trcrtegiufiir getni.sc.ht-~citz=~'cihligr
O~~tirrrirri~ng.~pro
blerne
LRN
Thesis,
University of Dortmund
rnit rwicihler Ditnen.\ioti DiplomaTEAM
Further reading
329
Further reading
More extensive treatments of issues related to the Baldwin effect can be found
in the literature cited in section C3.4.1. The following are notable foundation
TEAM LRN
and review papers.
330
Other oDerators
7. Sober E 1994 The adaptive advantage of learning and a priori prejudice From U
Biologicd Point of Vieuv: Essay in Evolutionary Philosophy (a collection of essays
by E Sober) (Cambridge: Cambridge University Press) pp 50-70
TEAM LRN
Index, Volume 1
A
Actuator placement on space structures
7
Adaptation 37
Adaptation in Natural and Artijicial
Systems (book) 46
Adaptive behavior 110
Adaptive landscape 36
Adaptive mutation 70
Adaptive Systems Workshop 46
Adaptive topography 24
Air combat maneuvering 9
Airborne pollution 8
Aircraft design 7
Alleles 64, 70, 164, 209, 263, 310, 31 1
Amino acids 33
Animats 1 10
ARGOT 77, 146
Arithmetic crossover 272
Artificial intelligence (AI) 90, 97, 189,
32 1
Artificial life (AL) 2, 321
Artificial neural networks. See Neural
networks
Autoadaptation 43
Automatic control 43, 94
Automatic programming 40
Automatically defined functions (ADFs)
110, 158
Autonomous vehicle controller 8
C
Canonical genetic algorithms I32
Cellular automata (CAs) 7
CHC algorithm 67
Chemistry 7
Chromatids 32
Chromosomes 27, 32, 33, 35, 64
Classification, applications 9, I0
Classifier systems (CFS) 2, 46
Clonal selection theory 37
Combinatorial problems (CES) 76
Combined representations 130, 131
Competitive selection 203
Compress mutation operation 158
Computer-generated force (CGF) 99
Computer programs 103, 109, 1 10
Computer simulation 1
Constant learning 314, 315
Constant-velocity environments 3 15
Control applications 8
Control systems 98
TEAM LRN
33 1
332
Convergence-controlled variation (CCV)
70, 75
Convergence-controlled variation
hypothesis (CCVH) 70. 71
Convergence rate 240, 241
theory 49
Corridor model 241
Creeping random search method 50
Criminal suspects 8
Critical learning period model 315
Crossover 64.65, 68-75, 235
bias 268, 269
in tetrad 32
mathematical characterizations 261,
262
mechanisms 256-6 1
one-point 69
points 257
probability 260
rate 257, 258
two-point 69
uniform 69
Crossover operators 76, 132. 238, 257,
258
characterizations 36 1
Cyanide production 3 1
Cyanogenic hybrid 3 I
Cycle crossover I45
Indes
Discount parameter 1 16
Discrete recombination 270
Disruption analysis 264
Disruptive selection 202
Distribution bias 269
Diversity 192
DNA (deoxyribonucleic acid) 33
Document retrieval 9
Domain-specific knowledge 3 18
Double helix 33
Drift 36, 46, 209
Drug design 98
Dynamic programming (DP) I16
Economics 7, 9
interaction modeling 7
Electrornagnetics 8
Elitist strategy 66, 210
Embryonic development 1 10
Encapsulate operator 158
Endosymbiotic systems 37
Engineering applications 7
Enzymes 33
Epistasis 31, 32
Equivalence 2 14- 18
Euclidean search spaces 81
Eukaryotic cell 33
Evaluations 89
D
noise 2 2 2 4
Darwinian evolution 89, 309
Evolution and learning 308, 309
Darwinism 27
Evolutionary algorithms (EAs) 6, 7,
Deception 72
20-2, 318
Deceptive functions 147, I49
admissible 191
Deceptive problems 72, 73
basic 59
Decision making, multicriterion. See
Boltzmann 195
Multicriterion decision making
common properties 59
Decision variables I27
computational power 320
Defining length 265
development 4 1
Delta coding 77
general outline 59
Deoxyribonucleic acid (DNA) 33
mainstream instances 59
Derivative methods 103-13
strict 19 1
Design applications 6, 7
theory 40, 41
Deterministic hill climbing 1
.SPY d.so specific types and
Dihybrid ratio 31
applications
Dimenxionality 2 1
Evolutionary computation (EC)
Diplodic representation 25 I
advantages (and disadvantages) 20--2
Diploid 27, 35
applications 4- 19
TEAM LRN consensus for name 41
Diploid representations 164
Index
333
discussion 3
population-based 83
history 40-58
steady-state 83
use of term 1
two-membered 38, 50
Eidutioiicrn' Computcition (journal) 47
Exons 33
Evolutionary game theory 37
Expected, infinite-horizon discounted
Evolutionary operation (EVOP) 40
cost 116
Evolutionary processes 37
Expression process 23 1
overview 23-6
Extradimensional bypass thesis 320
principles of 23-6
Evolutionary programming (EP) 1, 60,
136, 163, 167, 217, 218
F
basic concepts 89- 102
Fault diagnosis 7
basic paradigm 94
Feedback networks 97
continuous 95
Feedforward networks 97
convergence properties 100
Fertility factor 192
current directions 97-100
Fertility rate 192
diversification 44
Filters, design 6
early foundations 9 2 4
Financial decision making 9
early versions 95
Finite impulse response (FIR) filters 6
extensions 94-7
Finite-length alphabet 9 1
future research 100
Finite-state machines 33, 60, 9 I , 92, 95,
genesis 90
129, 134. 152, 153. 162, 236-8
history 40, 41, 90-7
Finite-state representations IS 1-4
main components 89
applications 1 5 2 3
main variants of basic paradigm 95
Fitness-hascd scan 273
medical applications 98
Fitness criterion 228
original 95, 96
Fitness evaluation 108
original definition 91
Fitness function 178
overview 40, 41
Fitness functions 172-5
self-adaptive 95, 96
monotonic 190, 191
standard form 89
scaling 66
v. GAS 90
strictly inonotonic 190, I9 I
Evolutionary robotics, see a l s o Robots
Fitness landscapes 229, 308, 31 1
Evolutionary strategies (ESs) 1, 48-5 1,
Fitness measure 235
60, 64, 81-8, 136, 163
Fitness proportional selection (FPS) 2 18
(1
I ) 48, 83
Fitness scaling 174. 175, 187
( 1 + A ) 48
Fitness values 59, 63, 66
(p
1 ) 83
Fixed selection 3 13, 3 15
( p A ) 48, 67, 83, 86, 167, 169,
Flat plate 48
189, 206, 210, 217, 220, 224, 230
Fouirclcttioris of' Getirtic Alpwithitis
( p , A) 189, 206, 210, 220, 222, 231
(FOGA) (workshop) 47
( P CL) 170
Functions 103, 105
alternative method to control internal
Fundamental theorem of genetic
parameters 86
algorithms 177
archetype 81, 82
Fuzzy logic systems 163
contemporary 83-6, 85
Fuzzy neural networks 97
development 40
Fuzzy systems 33
multimembered ( p > I ) 48
nested 86, 87
overview 48-5 1
TEAM LRN
Inde 5
334
G
Game playing 9
programs 45
Game theory 98
Gametes 27
Gametogenesis 27
Gaming 43
GauB-Seidel-like optimization strategy
87
Gaussian distribution 242
Gaussian mutations 24 1
Gaussian selection 3 I3
Geiringers theorem I1 263, 264
Gene duplication and deletion 3 I9
basic motivations 3 19
engineering applications 3 19, 320
formal description 32 1-3
historical review 3 19-2 I
Gene flow 36
Gene frequencies 36
Generation gap methods 205-1 1
historical perspective 206
Generational EAs 207
Generational models 2 16
Generic control problem 1 14
Genes 30, 33, 34, 64, 310
segregating independently 30
GENESIS 31 I
Genetic algorithms (GAS) 1. 2, 59, 60,
64-80, 103, 136, 155, 167
basics 65-8
breeder 67
canonical 64
generational 67
history 40-58
implementation 46, 65
messy 72, 73, 164
operation 70
overview 44-8, 64-80
pseudocode 65
steady-state 67
v. EP 90
see ctlso specific applications
Genetic drift 36, 46, 209
Genetic operators 65, 106-8, 110
Genetic Program Builder (GLIB) 158
Genetic programming (GP) 23, 60, 156,
167
defined 103-8
H
Hamiltonian circuit 139
Hamming cliffs 128
Hamming distance 75, 133, 148
Haploid 27, 35
Hardy--Weinbeg theorem 35
Heating, ventilation and air conditioning
( HVAC) controllers 98
Heterozygote 30
Heuristic crossover 273
H-infinity optimal controllers 8
Hinton and Nowlans model 310, 31 I .
312
Hitchhiking effects 7 1
Homozygotes 30
Hybrid algorithms 308-1 1
Hybridizations 60
Hydrodynamics 8 I
Hyerplanes 72, 177, 178, 179
analysis 69
Identification applications 7, 8
IEEE World Cotigress on Cotnputatioiiul
Intelligence ( WCCI) 4 1
Image processing 9
applications 44
Implicit parallelism 134, 190, 191
TEAM LRN
Incremental models 2 17, 220-2
Index
Inductive bias 268
Infinite impulse response (IIR) filters 6
Information retrieval (IR) systems 9
Information storage 9
Inheritance systems 28, 29, 37
Initialization 74, 89
Insert operator 245, 246
Intelligent behavior 42
Interactive evolution (IE) 228-34
application areas 231, 232
approach 229-3 1
difficulties encountered 23 1
formulation of algorithm 230
further developments and
perspectives 232, 233
history and prospects 228, 229
minimum requirement 228
overview 231
problem definition 229
samples of evolved objects 233
selection 229
standard system 229
Interceptor 94
Interdisciplinary Workshop in Adaptive
Systems 46
Intermediate recombination 85
Interrratiortal Coi2ferenc-e on Gerzetic
Algorithms (ICGA) 47
International Society for Genetic
Algorithms (ISGA) 47
Introns 163, 164
Inverse permutation 14 I , 142
Inversion 145
Inversion operator 77
Island models 36, 77
335
L
Lamarckian inheritance 308
Languages 60
Laplace-distributed mutation 240
Learning
and evolution 308, 309
as phenotypic variance 313. 314
Learning algorithms 308
Learning classifier systems (LCSs)
1 14-23
introduction 117, 118
Michigan approach 118
operational cycle 1 18
Pitt approach 1 I 8
stimulus response 1 18
structure 117
Learning models 316
Learning problems 1 14-1 8
Life cycle model 28
LINAC. See Linear accelerator
Linear accelerator, design 7
Linear-quadratic-Gaussian controllers
CQG) 8
Linear ranking 188, 21 5
Linguistics 9
Linkage, equilibrium 264
LISP 75, 108. 109, 128, 129
Local search (LS) 149
Machine intelligence 2
Machine learning (ML) problems 321
Mapping function I42
Markov decision problem 115
Mask 257
Mathematical analyses 44
Job shop scheduling (JSS) 5
Matrix representations 143-5
Joint plate 48
Maximum independent set problem 132
Juxtapositional phase 74
Medical applications 98
Medicine 44
Meiosis 27, 32
Memory cache replacement policies 5
Mendel, Gregor 28
k-ary alphabet 134
Mendelian experiment 28, 29, 30
Knapsack problems (KP) 6, 132
Knowledge-augmented operators 3 17-1 9
Mendelian inheritance 261
Mendelian ratios 31
Kohonen feature map design 6
Messenger RNA (mRNA) 33, 34
Messy GA 251
TEAM LRN
336
Messy genetic algorithms (mGAs) 72,
73, I64
Metropolis algorithm 196
Military applications 9
Minimax optimization 87
Mi x ed - integer opt i mi zat i on 87
Mixed-integer representation 163
Mixing events 268
Monohybrid ratio 30
Monte Carlo (MC) generators 7
Multicriterion problems 5 0
Multimodal objective functions 84
Multiple criterion decision making
(MCDM) 22
Multiple-input, multiple-output (MIMO)
model 98, 99
Multipoint crossover 266
Multiprocessor system 5
Multivariate Lero-mean Guassian
random variable 137
Mutation 23. 35, 42, 59, 60, 61, 68-75,
89, 108, 152, 228, 237-55
?-opt, .?-opt and k-opt 244, 245
function 239
optimum standard deviations 24 1
successful 2 1 I
Mutation function 3 I4
Mutation operators 84,92, 93. 125, 132,
158. 237-5s
Mutation-selection equilibrium variance
3 14
Mutations 36
Inde?
Nonlinear optimimtion problems I
Nonlinear ranking 188. 189, 2 I5
Nonlinear ranking selection 202. 203
Nonoverlapping populations 205
Nonoverlapping systems 208
Nonregressive evolution 43
Nonuniforni mutation 243
n-point crossover 259. 266
0
Object parameters I36
Object variables 136-8
Objective functions 172, 173, 178
Objective values 192
Offspring machines 42, 152
One-point crossover 258, 265, 266, 271
Online planning/navigating 5
Operator descriptions 149
Operon system 34
Optical character recognition (OCR) 9
Optimimtion methods I , 2, 89, 160
Optimum convergence 24 1
Order-based mutation 246
Ordering schemata (o-schemata) 146.
147
Oren-Luenberger class 85
Ovcrlapping populations 205, 208, 209,
2 10
Ovum 27
Packing problems 6
Pairwise implementation 70
N
Pairwise mating 70, 71. 75
Near misses 308
Parallel genetic algorithms (PGAs) 77
Neighborhood, model 77
PLIrct llel PmhIein Sol 1siirg j mm Ncrtii re
Neo-Darwinism 23, 24, 37
( P P S N ) (workshop) 41
Nesting technique 86
Parallel recombinative simulated
Network design problems 6
annealing (PRSA) 196, 197
Neural networks 6, 43, 44, 99, 163
parameters and their settings 197
design 97
pseudocode for common variation
training 43, 44
I97
Neutral molecular evolution theory 37
working mechanism 196, 197
Niching methods 50
Parameter optimiation 133
No-free-lunch (NFL) theorem 20, 2 I
Parameter settings 22
Nodes 104
Parasites 120, 121
Nondetermi nis t ic-polynomial- ti me ( NP)
Parse tree representation 155-9
complete problems 97
complex numerical function I57
TEAM LRN primitiite language 156, 157
Nondisruptive crossovers 267
Index
337
Q
Q-learning 116
Q-values I16
Quantitative genetics models 3 13-16
Query construction 10
Random decisions 48
Random keys 146
Random mutation hill climbing
(RMHC) 243
Random program trees 106
Random search 1
Randomness 1
Rank-based selection 66, 169, 187-94.
209
overview 187, 188
theory 1 9 0 4
Ranking 187, 188, 215, 216, 218, 219
Real-valued parameters 75
Real-valued vectors 60, 128, 134,
136-8, 239-43, 270-4
Recombination 59, 60, 85, 106, 107,
152, 228, 256-307
dynamics 262, 263
events 257
formal analysis 261
Recombination bias 269
Recombination distributions 261, 264,
265, 270
Recombination events 265, 266, 269
Recombination-mutation-selection loop
61
Recurrence relations 263
Reduced surrogate recombination 26 1
Reinforcement learning problem 1 15.
I17
Replacement
biased v. unbiased strategy 67
selection 67
Representation 75-7, 127, 128, 228. 235
alternative 145, 146
guidelines for suitable encoding
160-2
importance of 128-30
nonstandard 163, 164
see c i l s o specific representations
Representations 127-3 1
Reproduction 66
Reproductive plans 45
TEAM LRN
Reproductive rate I92
338
Resource scheduling I39
REVOP 40
Ribonucleic acid (RNA) 33
Robbins equilibrium 264
Robots
applications 9
control systems 8
optimization applications of EP 43
path planning 5
.see also Evolutionary robotics
Robustness 1, 2, 20
Roulette wheel, sampling algorithm 175,
176
Route planning 43
Routing problems 4, 5, 97
Rule-based learning, Pitt approach to 47
Inder
Index
339
Supersolution 168
Superstrings 199
Supervised learning 1 14, I 15
Swap mutation 245, 246
Switch mutation operator 248
System identification problems 8
T
Takeover time 179, 180, 182, 193
Target sampling rates 177, 190
Terminals 104
Termination criterion 6 1
Testing applications 7
Tetraploid 35
Three-dimensional convergent-divergent
nozzle 49
Threshold 189
Threshold selection 189, 190
Timetables 5
Tournament competition 97
Tournament selection 66, 18 1-6, 20 1,
202, 215
binary 168
concatenation of tournaments 183
formal description 182
loss of diversity 184
operator 168
parameter settings 182
properties 183-5
working mechanism 181, 182
Tournament size 181, 184
Transcription process 33, 34
Transfer RNA (tRNA) 34
Translation process 34
Transmission probabilities for
second-order schemata 268
Transportation problem 4, 5
Traveling salesman problem (TSP) 4,
43, 76, 97, 139, 143, 146, 147,
161, 162, 240, 245, 318
Tree-like hierarchies 75
Tree-structured genetic material 107
Tree-structured representation 104
Truncation selection 189, 217, 218
Two-dimensional foldable plate 8 1
Two-dimensional packing problems 6
Two-membered ES 48, 50
Two-phase nozzle optimization 49
Two-point crossover 267
U
Uncertainty 3
Unequal-area facility layout problem 6
Uniform crossover 73, 85, 259, 267
Uniform scan operator 271
V
Variable lifespan 203
Variable-length genotypes 3 19, 320
Variance 208, 209
Variation 89
Vehicle routing problem 4, 5
Virtual machine design task I05
W
Watson-Crick model 33
Z
Zero-one knapsack problems 6
Zygote 35
TEAM LRN