Professional Documents
Culture Documents
Indexing terms: Adaptive control systems, Biomedical control, Evolutionary adaptation, Furzy logic control, Genetic algorithms, Machine learning
There is to-date no generalised method for the formulation of fuzzy control strategies, and design remains an ad
hoc trial and error exercise.
This paper discusses a new, objective approach to the
design of fuzzy controllers. The approach uses genetic
algorithms, a recent search and optimisation technique
[4,51, to optimise the parameters of a fuzzy controller. A
similar approach was recently reported in References 6
and 7 in which the method was used for the optimisation
of the membership functions of a fuzzy control rule-base.
The present study, however, extends the method to the
complete process of design of a multivariable controller
for a nonlinear biomedical process, namely, control of
anaesthesia.
2.1 Fuzzy design methodology
introduction
Simulate process
cesign
Fig. 1
Genetic algorithms (GA) are exploratory search and optimisation procedures that were devised on the principles
of natural evolution and population genetics. The basic
concepts of GA were developed by Holland [22,23], and
subsequently in several research studies. Goldberg [5]
and Davis [4] provide recent comprehensive overviews
and introductions to GA.
There are several differences between the functioning
of C A and traditional optimisation techniques, especially
those based on gradient methods. These differences
include the following:
(i) GA work on a coding of the parameters to be optimised, rather than the parameters themselves. Binary
coding is normally used and has been suggested to be
optimal in certain cases, however this is neither a
restriction nor a requirement of GA.
(ii) GA search a space using a population of trials,
representing possible solutions to the problem. The initial
population, usually, consists of randomly generated
individuals.
(iii) GA use an objective function assessment, to guide
the search of the problem space.
(iv) GA use probabilistic rules to make decisions.
Typically, the GA starts with little or no knowledge of
the correct solution depending entirely on responses from
an interacting environment and its evolution operators to
arrive at good solutions. By dealing with several independent points the GA samples the search space in parallel,
and hence is less susceptible to getting trapped on suboptimal solutions. In this way, GA have been shown to
be capable of locating high performance areas in complex
domains without experiencing problems with high
dimensionality, or false peaks, as would occur with hill
climbing optimisation techniques.
As has been stated, GA imitate natural evolution, and
hence include operations such as reproduction, crossover, mutation and inversion. A simple GA has four
features; Population size, Reproduction, Crossover, and
Mutation. Reproduction is a process in which a new generation of population is formed by randomly selecting
strings from an existing population, according to their
fitness. This process results in individuals with higher
fitness values obtaining one or more copies in the next
generation, while low fitness individuals may have none;
it is for this reason called a survival of the fittest test.
Crossover is the most dominant operator in a CA, and is
responsible for producing new trial solutions. Under this
operation, two strings are selected to produce new offspring by exchanging portions of their structures. The
offspring may then replace weaker individuals in the
I E E Proc.-Control Theory Appl., Vol. 142, N o . 3, M a y 1995
Fig. 2 illustrates a typical coding format for a population of fuzzy rule-bases comprising two conditions E
and CE, and one action U.This is the structure of rules
adopted in many fuzzy control applications, where E
I
rule 8
31-29
28-24
Fig. 2
En = ,y ui = [d,U * , . . ., U]
A =
. . . uF, U:. . . urn]
(4)
where U;, or U;. are the codes of the ith condition or
action fuzzy sets.
164
rule 1
Fig. 3
= Ami,
+ (Amx
A,JA(u)
(5)
where 0,is the initial number of schemata in a population, 0,is number of schemata in a population at convergence, and T, the time taken to reach convergence.
The number of different schemata in a converged population is the number of different schemata represented by
a single string, and is given by 2, where I is the string
length. For example, a 2-bit binary string {lo} can represent the schemata { 10, *0, 1*, **}. Thus, the number of
schemata in a population of N different individuals is
equal to 2. However, this is true for small size populations and the number increases to 3 for very large
populations when nearly all possible strings are represented in the population. Hence, the figure of merit
proposed above can be expressed as
[/A,
NM
Fig. 4
parameter within its bounding constraints. A ( d ) is dependent on the decimal value of U, Dec (U) E [O, 1, . . ., 2
- 11 and the cardinality of the coding, equal to 2 - 1
where, lj is the codes length. The decimal value of a code
is given by
D e c ( d )= ( i l u i 2 i - 1 )
(6)
eo- 0,
A=-
T,
IEE Proc.-Control Theory AppL, Vol. 142, No. 3, May 1995
A=-
2N - 2
nc
(9)
where N is the population size, nc is the number of generation to reach convergence, and t, is the processing time
per generation. A computer program was developed in
the study to maximise the figure of merit for different
values of 1. A surprising result was that although the
required population setting increased linearly when the
time to reach convergence was fixed, it was found to be
constant regardless of the string length when the time to
reach convergence is allowed to vary. Perhaps more surprising was the fact that this population size turns out to
be equal to 3.
The above study provides a sound basis for selection
of a suitable population size, although it ought to be said
that the suggested population size of 3 should not be
taken rigidly. It happens to be the minimum number that
can allow selection of different pairs of individuals for
recombination. However, a population size of 3 does not
provide any meaningful diversity in the population.
Hence, there is a need to increase the population size to
achieve a good compromise between diversity and convergence time. In the current study, population sizes
between 20 and 40 were used. Although these could be
regarded as very small sizes in view of the string lengths
(typically 2080 bits) it has been found that even the smallest size achieved good results with suitable recombination operators. This result is supported by the work of
De Jong and Spears [32] and Syswerda [35] who established that recombination operators that are highly disruptive to schemata do actually achieve better
performance than the traditional one-point crossover in
GAS using small population sizes. In the study by Goldberg [31] it was suggested that the rapid convergence of
small populations can be used to advantage to improve
ofline performance, by reinitialisation of the population
when it becomes uniform. The GA in the present study
does not reinitialise populations but permits high rates of
disruption to existing populations, through increased
mutation.
Traditionally, the
number of cross-over points used in a GA, which determines the number of segments of the parent codes that
are exchanged, has been set to 1. The supporting basis for
this choice is the theoretical result expressed by the
schema theorem. Specifically. the one-point crossover
produces the minimal disruption to schema allocation.
However, there have been experimental studies which
have suggested the use of more crossovers and higher
rates of mutation. The study carried out by Syswerda
[35], for example, indicated that more disruptive forms of
mating can improve the performance of a CA. A disruptive mating process is one which results in offspring that
are significantly different from their parents. Disruption
is considered advantageous when the population
becomes homogeneous, when it can improve the productivity of a C A by continued creation of new trials.
Under these circumstances, more conservative operators,
such as the one-point crossover, produce offspring that
are identical to !he parents and, hence, the search stagnates. A latter study by De Jong and Spears [32] established that the improvement in performance was more
with smaller population sizes than larger ones.
There are two types of multiple crossover, namely
n-point crossover and uniform crossover [37]. In n-point
crossover, n breakpoints are chosen at random at which
points the parents bits are alternately passed on to the
offspring. Offspring take genes from one of the parents
until a crossover is encountered, at which point they
switch and take genes from the other parent. Another
crossover operator, called punctuated equilibria. functions in a very similar way except that the crossover
points are themselves coded in a separate chromosome,
which is also subject to genetic operators. In uniform
crossover there is a probability p , that the offspring will
take a bit from the corresponding positions of either
parent. For example, when pc is 0.7 an offspring will
acquire the genes of one of the parents with a probability
of 0.7 (and the genes of the second parent with a probability of 0.3). A value of p , equal to 0.5 thus implies
random crossover since the offspring take on bits from
the parents with equal likelihood, while p , equal to 1.0 or
0.0 creates offspring that are clones of the parents. The
mechanisms of both types of crossover are illustrated in
Fig. 5.
breok-points
t = e x c F n g e c bits
Fig. 5
il
(n = 3tpoint crossover
h uniiarm crossover
[ .J = & . -J
&m m
(13)
where cminis a base-line value of evaluation. Often E,,,;" is
not known a priori and the popular strategy is to continually assign it values which have been observed in previous generations. For example, if all individuals in a
population have evaluations in the range C7.1 ... 9.23 a
fitness base-line could be set at 6.0, and hence, by subtracting this number from all evaluations, new fitness
values will be in the range [1.1 ... 3.21. The best individual now has a three times chance of selection over the
worst.
Functional normalisation is a new technique considered
in this study. One of the most common fitness normalisation techniques is linear normalisation [SI, which
scales the objective evaluations according to the equation
i j = ai+ b
(14)
The coefficients a and b are chosen to ensure that both
the original objective evaluation and the scaled fitness
have the same mean, and that the maximum value of the
scaled fitness is a certain multiple of this mean. One
shortcoming of the technique is that it does sometimes
result in negative fitness values and it is, therefore, necessary to take precautions to deal with this. Functional
normalisation, on the other hand, 'stretches' or 'compresses' the set of evaluations to maintain a given ratio of
minimum to maximum fitness values, which is called the
selection pressure K. This is done by calculating new
maximum and minimum fitness values to give the desired
ratio in each generation. The intermediate objective
evaluations can then be interpolated linearly, or nonlin168
early. Linear transformation of a set of objective evaluations, for example, is carried out according to the
following steps:
1 Calculate the new range of fitness values to obtain
the required pressure, K,
where
and E,, are observed minimum and maximum
evaluations and K is a specified value of the selection
pressure required
Kimin
(16)
2 Calculate the slope of the (linear) fitness transformation
imax
Er=,
(fuzzy controller)
(cost function)
learning model
(senetic alqorithm)
I
Fig. 6
populotion model
(trials)
fuzzy controller implementable with the trial knowledgebase. The basic control objective is simply stated as the
ability to follow a setpoint with minimal error. This
objective can thus be expressed in terms of minimisation
of controller performance indices which are in common
use. These include integral of absolute errors (IAE),
integral of square errors (ISE) and integral of time
multiplied absolute errors (ITAE). Each of these indices
has its own merits. For example, ITAE penalises errors at
large values of time and leads to reduction in steady state
errors at the expense of transient errors, while ISE is a
more suitable mathematical analysis criterion. Furthermore, the GA is only able to optimise the characteristics
explicit in the cost function. In this work, it was found
necessary to incorporate a penalty of excessive control
effort into the evaluation cost function, in particular to
prevent relay-type control. Thus, an index called integral
of absolute control rate and error (IACRE) was derived.
A general mathematical equation for such an evaluation
cost function is
7
=
k = l
(19)
p . = i=
6;
(20)
There are other factors which could improve the performance of a GA, especially for fuzzy controller design.
These include: the need to speed up the evaluation using
parallel processing, the need to reduce the size of the
rule-bases, and improvement of the learning process with
knowledge-based genetic operators. The time taken in
evaluation of genetic structures imposes restriction on the
size of the population and also the number of generations
required to run the GA to a final solution. To alleviate
the problem, parallel processing can be used to reduce
generational execution times. There are two common
approaches to parallel processing in GAS. The first
approach is a one-population GA which only uses the
multiple processors to divide the evaluation task. Each
processor evaluates a given number of individuals and
returns their objective assessment to the population. The
second alternative is to have separate subpopulations on
each processor which develop individual solutions. In
such cases, usually, there is occasional migration of
members between the subpopulations. This model of
GAS is commonly known as a parallel or distributed GA
(PGA) [41, 421. Although studies have shown this mode
of GA to yield better solutions than the one-population
GA, it does not solve the problem that is of concern to
this study: reducing execution times. Thus. the first
method of parallelisation is preferred, because the
reduction times takes precedence over any possible
achievement of an optimal solution with a PGA.
Human reasoning and decision making are guided by
knowledge. Furthermore, fuzzy logic systems are
designed to imitate this reasoning mechanism and are
characterised by rules that are expressive knowledgebased models of the task domain. In contrast, C A in their
basic form are dependent only on coding, function evaluations and exchanges of the codes between individuals.
This is both an advantage and a disadvantage. Although
the ability to work without problem specific information
allows GAS to be applicable to different problems
without specific tailoring, failure to make use of available
domain knowledge would definitely put them at a disadvantage against other methods that could use such
information. One way of utilising available knowledge
would be to use it to seed the initial populations. The
advantage of this is that it can result in a reduction in the
number of generations required to reach a satisfactory
solution. Another way of utilising available knowledge is
to use knowledge-based genetic operators that exploit
any known regularities in the domain space. The evaluation of a genetic structure representing a fuzzy rule is
carried out at a knowledge level since the structures are
first interpreted into fuzzy rules, and then applied to the
169
[;I=[
(1
1.&-(1 + 10.64s)
+ 3.08sX1 + 4.81s)(l + 34.42s)
(1
+ 2.0s)
crossover point
I
c L!d] e
[ A
Fig. 7
a I b
IC I
D ] e I f
parents
offspring
1 ID]
A l B I C I d ] E I F
F ]
Knowledge-based crossover
F are code segments represenling linguistic names and fuzzy sets of rules
a .. . f and A . . .
Fig. 8
Knowledge-bused mufation
Mutation changes the linguistic fuzzy set to another. a distance of one or two
away
usual application of 'generational replacement' the alltime-best individual may not exist in all generations since
entire populations are replaced in each sample step. This
is evidenced by the fact that the best individual in the
latter generations is worse than the all-time-best.
Without any means of deterministically selecting or
120
01
'_r(
1201
I
1 @@OO
number of evaluations
200@0
Fig. 9
Learning profile of G R method
0 NORM
RANK
WIND
~1 mean fitness
h best fitness
01
1@00@
number of evaluations
?@Cl 30
Fig. 10
0 NORM
A RANK
+ WIND
a mean fitness
h best fitness
Best values far b : 0 27.31; A 27.45;
+ 27.80
120
4 2 Performance results
The results of the learning process are the rule bases that
are used to control the multivariable anaesthesia model.
The rules obtained with the steady-state and the selective
breeding reproduction methods together with the ranking
I
120
I
1
01
20000
120
10000
number of evaluations
20000
Fig. 11
Fig. 12
10000 0
number of evaluations
b
A RANK
+ WIND
a mean fitness
b best fitness
Best values for b : 0 27.25; A 26.84;
172
0 NORM
+ 26.14
0.6
Lo[
Error
PB
PM
PS
PO
NO
NS
NM
NB
Another way of looking at the rules is a control decisions' table that shows inferred actions for different
process states (error and change-in-error).An example of
NB
NM
NS
NO
PO
PB
PO
PO
PS
NS
NS
PO
PS
PS
PM
NS
NB
NO
NM
PB
PO
PS
PO
NS
NM
NB
NB
PO
PB
PS
PM
NS
NS
NS
NS
PS PB
PO PO
PS PB
PO PS
NS NB
NS NS
N M NS
NO NO
NS
NO
PS
PM
PB
PO
PB
PS
PM
NB
NS
NO
NS
PO
PB
PM
PB
NO
NB
NO
NB
~_______
h
:
50 100 150 200
0
0
'
50
minutes
100 150
minu,tei
200
Fig. 14
Multivariable response, from selective breeding and fitness
normalisation methods
MAP response
band b MR responses. showing interactions from MAP
n and a'
10
0.8
0.6
d 40
I?
'h
6 0.4
20
E
0.2
'0
Fig. 15
methods
OCE
-E
100
150
200
Z 0 L L L - L 0
50
100
minutes
b
50
150
200
minutes
b'
o and a'
MAP response
band b' MR responm, showing interactions from MAP
CE
E
-10
-8
-6
9.8
8.3
5.7
5.1
-1.6
-0.7
-2.0
-4.8
-3.9
-3.5
-3.5
9.8
8.6
6.6
5.1
-1.6
-0.8
-2.5
-4.9
-3.9
-3.5
-3.5
9.4
8.6
7.2
4.4
-1.4
-0.7
-1.9
-5.0
-4.3
-3.0
-3.0
-4
-2
7.9
7.0
5.2
3.5
1.9
-0.0
-0.0
-2.5
-3.6
-3.3
-3.3
5.1
5.2
4.9
2.0
1.1
0.1
1.1
-1.3
-3.5
-4.0
-3.9
5.5
5.2
5.1
1.0
1.2
0.7
0.8
-3.0
-2.9
-2.8
-3.5
4.5
4.5
4.0
1.9
1.5
-1.2
-2.9
-0.5
0.9
10
~~
10
8
6
4
2
0
-2
-4
-6
-8
-10
6.1
6.1
4.6
-1.0
-0.6
-0.4
-1.3
-3.9
-3.5
-3.1
-3.1
6.1
6.2
4.6
3.2
1.2
-0.6
-2.9
-4.2
-4.7
-3.4
-3.1
-0.8
-3.0
2.9
1.9
2.9
2.2
2.7
2.5
1.3
1.3
1.5
1.4
0.2
2.6
0.7
2.7
1.8
1.7
1.8
1.8
-2.7 -2.7
-4.7 -4.7
173
control decision tables derived from the linguistic rulebases is given in Table 2. This table is obtained from the
rule-base by scanning the input space and inferring an
input for each point. The table is similar to 'look-up
tables' sometimes used for direct fuzzy control [47,48].
+-a'4 0
L
E
a'
100
60
2
rr' LO
I
20
minutes
minutes
b
b'
normnl.
Fig,
response,from steady
isation methods
(1 and a' MAP response
b and b MR responses,showing interactions from MAP
=t
"
'
"
'
rrinuies
b
b'
Fig. 18 Multivariable response, from a generational gap GA
a and a' MAP response
band b MR responses, showing interactions from MAP
For b : ITAE = 6525.92; IAECE = 90.60
40
01
minutes
a'
Conclusion
'
174
References
9th international joint conference on Artificial intelligence (MorganKauffman Publishers, 1985). pp. 573-580
20 KARR, C.L.: Design of an adaptive fuzzy logic controller using a
genetic algorithm, in SCHAFFER, I. (FA.): Proceedings of the third
international conference on Genetic algorithms (Morgan Kauffman
Publishers, Cambridge, MA, 1989), pp. 450-457
21 THRIFT, P.: Fuzzy logic synthesis with genetic algorithms, in
BELEW, R., and BOOKER, L. (Eds.): Proceedings of the fourth
international conference on Genetic algorithms (Morgan Kauffman
Publishers, Cambridge, MA, 19911, pp. 509-513
22 HOLLAND, J.H.: &Geneticalgorithms and the optimal allocation of
trials, SIAM J. Compur., 1973,2, (2), pp. 89-104
23 HOLLAND, J.H.: Adaptation in natural and articifical systems
(Addison-Wesley, 1975)
24 ANTONISSE, H.J., and KELLER, K.S.: Genetic operators for high
level knowledge representation, in GREFENSTETTE. J. (Ed.): Proceedings of the second international conference on Genetic algorithms (Lawrence Erlbaum Publishers, Hillsdale, NI. 1987). pp.
69-76
25 CARUANA, R.A., and SCHAFFER, J.D.: Representation and
hidden bias: Gray vs. Binary coding for genetic algorithms, in
LAIRD, J. (Ed.): Proceedings of the fifth international conference on
Machine learning (Morgan-Kauffman Publishers, Camhridge, MA.
1988), pp. 153-161
26 SCHRAUDOLPH, N.N., and BELEW, R.K.: Dynamic parameters
encoding for genetic algorithms, Machine Learning, 9 (Kluwer Academic, 1992), pp. 9-21
27 GREFENSTETTE, J.J.: Optimization of control parameters for
genetic algorithms, IEEE Trans. Syst. Man & Cyhern., 1986,
SMC-16, pp. 122-128
28 DAVIS, L.: Adapting operator probabilities in genetic algorithms,
in SCHAFFER, J. (Ed.): Proceedings of the third international conference on Generic algorithms (Morgan Kauffman Publishers, Cambridge, MA, 1989). pp. 61-69
29 SCHAFFER, J.D., CARUANA, R.A., ESHELMAN, L.J., and DAS,
R.: A study of control parameters affecting online performance of a
genetic algorithm for function optimization, in SCHAFFER, J.
(Ed.): Proceedings of the third international conference on Genetic
algorithms (Morgan Kauffman Publishers, Cambridge. MA, 1989),
pp. 51-60
30 DE JONG. K.A.: Are genetic algorithms function optunizers?, in
MANNER, R., and MANDERICK, B. (Eds.): Proceedings of the
second workshop on Parallel problem soloing Jrom nature (SpringerVerlag, Berlin, 1992). pp. 2-13
31 GOLDBERG, D.E.: Sizing populations for serial and parallel
genetic algorithms, in SCHAFFER, J. (Ed.): Proceedings of the
third international conference on Genetic algorithms (Morgan KauITman Publishers, Cambridge, MA, 1989),pp. 71-79
32 DE JONG. K.A., and SPEARS, W.M.: An analysis of the interacting roles of population size and crossover in genetic algorithms,
in SCHWEFEL, H.-P., and MANNER, R. (Eds.): Proceedings of
the first workshop on Parallel problem solving Jrom nature,
Dortmund, West Germany (Springer-Verlag, Berlin. 1990), pp,
39-47
33 ODETAYO, M.: Optimal population sue for genetic algorithms
an investigation. IEE digest of the colloquium on genetic algorithms in control systems engineering (London, May 1993)
34 ROBERTSON, G.G.: Population size in classifier systems, in
LAIRD, J. (Ed.): Proceedings of the fifth international conference on
Machine learning (Morgan-Kauffman Publishers, Cambridge, MA,
1988), pp. 142-152
35 SYSWERDA, G.: Uniform crossover in genetic algorithms, in
SCHAFFER, I. (Ed.): Proceedings of the third international conference on Genetic algorithms (Morgan Kauffman Publishers, Cambridge, MA, 1989). pp. 2-9
36 SYSWERDA, G.: A study of reproduction in generational and
steady state genetic algorithms, in RAWLINS, G.J.E. (Ed.): Proceedings of a workshop on foundations of Generic algorithm and
classifier systems (Lawrence Erlbaum Publishers, Hillsdale, NJ,
1991), pp. 94-101
37 SCHAFFER, J.D., and MORISHIMA, A.: An adaptive crossover
distribution mechanism for genetic algorithms, in GREFENSTETTE, 1. (Ed.): Proceedings of the second international conference on Genetic algorithms (Lawrence Erlbaum Publishers,
Hillsdale, NJ, 1987), pp. 36-40
38 MUHLENBEIN, H.: How genetic algorithms really work. I: Mutation and hill-climbing, in GREFENSTETTE, 1. (Ed.): Proceedings
of the second international conference on Genetic algorithm
(Lawrence Erlhaum Publishers, Hillsdale, NI, 1992). pp. 15-25
39 BACK, T.: Optimization of mutation rates in genetic search, in
FORREST, S. (Ed.): Proceedings of the fifth international conference on Genetic algorithms (Morgan Kauffman Publishers, C a n bridge, MA, 1993).pp. 2-8
175
176
45
46
47
48