Professional Documents
Culture Documents
SO965-9978(96)00011-7
the mean field theory (MFT), for finding approximate solutions to the
partitioning of the finite element meshesis investigated.The partitioning is
based on the recursive bisection approach. The method of mapping the mesh
bisection problem onto the neural network, the solution quality and the
convergence times are presented. All computational studies were carried out
using a single T800 transputer. Copyright 0 1996 Civil-Comp Limited and
Elsevier Science Limited
1 INTRODUCTION
Combinatorial optimization problems arise in many
areas of science and engineering. Unfortunately, due to
the NP (non-polynomial) nature of these problems, the
computations increase with the size of the problem.
Most computational methods that have so far been
developed which generally yield good solutions to these
problems rely on some form of heuristic. Artificial
neural networks (ANNs) make use of highly interconnected networks of simple neurons or processing
units which may be programmed to find approximate
solutions to these problems. They are also highly
parallel systems and have significant potential for
parallel hardware implementation.
The origin of the optimization neural networks
goes back to the work by Hopfield & Tank which
was a formulation of the travelling salesman problem
(TSP). The Hopfield network is a feedback-type of
neural network where the output(s) from a processing
unit is fed back as the input(s) of other units through
their interconnections. This type of network structure is
a nonlinear, continuous dynamic system. Figure 1
illustrates a typical feedback neural network.
Following the poor performance of Hopfield networks in determining valid solutions to the TSP
problem, there followed considerable research effort
to improve the performance of this type of network and
to find ways of applying it to other optimization
problems.
At about the same time of the emergency of Hopfield
networks, a new optimization method called simulated
annealingz,3 was researched and developed. This
104
0
-
- 0
0-
0
0
0
(1)
where S represents the state of the network, Z is the bias,
sN), N is the number of processing units,
S=(q,sz,...,
and tij represents the strength of the synaptic connection
between the units. It is assumed that the tij matrix is
symmetric and has no self-interaction (i.e. tji = 0).
In order to move E(S) downwards on the energy
landscape, the network state is modified asynchronously
from an initial state by updating each processing unit
according to the updating rule:
c 1
N
Si
sgn
tijSj
(2)
Zi
=I
where output of ith unit is fed to the input of the jth unit
by the connection tijs The symmetry of the matrix tij
with zero diagonal elements enables E(S) to decrease
monotonically with the updating rule.
0
Fig. 3. A simplified illustration of magnetic material described
by an Ising model.
Buffer
annealing
is a probabilistic
hill-climbing
105
(9)
and
aE(v)
x+Ui=O
I
shownin Fig. 3.
solutions with lower energy. The method carries out
neighbourhood searchesin order to find new configurations using the Boltzmann distribution:
e-wIT
h(S) = z
(3)
where T is the temperature of the system and Z is the
partition function of the form:
z=c(S)e-E(S)T
However simulated annealing involves a stochastic
search for generating new configurations. In order to
reach good solutions, a large number of configurations
may have to be searched, which involves the slow
lowering of the temperature and therefore is a very CPU
time-consuming process.
2.4 Mean field annealing network
C fi
j=l
lrn
--OO
dvj
Frn
&je-E'(v~u~T)
-i'x
(5)
-3.0
-2.0
-I .o
0.0
I .o
2.0
3.0
4.0
5.0
(6)
Fig. 5. The gain function for different temperatures.
106
(12)
(13)
107
decreased.
Figure 7 was the result of the bisection of the mesh
shown in Fig. 6, where the initial temperature was
chosen as 3. The total number of iterations for the
bisection was 25.
The effect of the decrease in temperature upon
individual spins for this example is represented in
Fig. 8. This figure shows that at high temperatures the
spin average diverges to near zero for all the elements
which indicates that the bisection is maximally disordered. As the temperature is decreased, the system
reaches a critical temperature where each element starts
to move significantly into one or another of two
bisections. At sufficiently low temperatures the spins
saturate at or near values of 1 or - 1. Therefore, if a net
is initialized with a temperature just above or equal to
the critical temperature, the fastest convergence to a
good global minimum should occur. Thus a bisection
for the mesh shown in Fig. 6 was carried out but this
time with an initial temperature 0.8. The total number of
iterations for this bisection was 11 which produced the
same result shown in Fig. 7.
The initial imbalance factor o is usually selectedas 1.0
for a balanced bisection. This value in most cases
ensuresa balanced bisection but a minimum cutsize may
not be produced. Therefore it was decided to carry out
an inverse annealing or, in other words, incrementation
of the initial value of (Y.This was carried out by selecting
a small initial value for IY (for example, 50.1) and once
the neural network optimization has started, CI is
increased by a factor (for example, 1.5) after one
complete iteration of eqn (13) until it reaches 1.0. The
108
l
l
A
Fig. 9. A squaredomain with in-plane load.
data consisted of three side lengths, three internal angles
and two scaled mesh parameters of each element. The
output data consisted of the number of generated
triangles in the refined mesh. A network was trained
and once a desired trained network is achieved it may be
used to predict the number of elements which may be
generated within an element of a coarse mesh.
4.2 The predictive mean field Hamiltonian
The original equation of the MFA mesh bisection
(i.e. eqn 13) was modified in order to accommodate the
predicted number of elements which may be generated
within an element of the coarse mesh. This equation was
thus modified and the MFA bisection equation for using
109
99
108
97
108
103
103
103
103
Diff.
%age diff.
-4
5
-6
5
-3.88
4.85
-5825
4.85
1
2
3
4
Required
elements
Diff.
%age diff.
103
103
103
103
-4
5
-6
5
-3.88
4.85
-5.825
4.85
Interfacing nodes
62
62
SGM
MFA
Time (min)
1.067
0.05
of
Fig. 11. The initial mesh with 46 elements divided into four
partitions using the SGM.
Fig. 13. The initial mesh with 46 elements divided into four
partitions using the MFA method.
110
Subdomain
Generated
Required
no.
elements (actual) elements
Diff.
%age diff.
1
2
3
4
-9.75
-1.75
-3.75
15.25
-5.78
-1.03
-2.22
9-03
159
167
165
184
168.75
168.75
168.75
168.75
kr
1
2
3
4
152
184
177
167
Required
elements
Diff.
%age diff.
Interfacing nodes
Time (min)
76
84
3.2
0.233
5.2 Example 2
This example is an L-shaped domain shown in Fig. 15,
which is uniformly meshed, and the result is an initial
coarse mesh with 126 elements, which is shown in
Fig. 16. This initial mesh was then decomposed using
the SGM and the MFA method.
112
Fig. 23. The initial mesh with 153 elements divided into eight partitions using the SGM.
113
144
152
146.5
146.5
149
146.5
150
147
146.5
146.5
-2.5
5.5
2.5
3.5
0.5
-1.707
3.75
1.707
2.39
0.341
Generated
Required Diff. %agediff.
Subdomain
no.
elements(actual) elements
1
136
146.5 -10.5 -7.16
2
160
146.5
13.5
9.21
3
4
5
6
7
8
151
157
146
138
141
137
146.5
146.5
146.5
146.5
146.5
146.5
4.5
10.5
-0.5
-8.5
-5.5
-9.5
3.07
7.16
-0.34
-5.80
-3.75
-6.48
Partitioning method
SGM
MFA
Interfacing nodes
179
188
Time (min)
4.267
0.333
5.3 Example 3
6 CONCLUDING
REMARKS
114
Fig. 25. The initial meshwith 153elementsdivided into eight partitions using the MFA method.
ACKNOWLEDGEMENTS
REFERENCES
1. Hopfield, J. J. & Tank, D. W. Neural computation of
decisions in optimization problems. Biol. Cybernetics,
1985,52, 141-152.
2. Kirkpatrick, S., Gelatt Jr, C. D. & Vecchi, M. P.
Optimization by simulated annealing. Science, 1983,
2204598, 671-680.
3.
Mesh partitioning
4. Simon, H. D. Partitioning of unstructured problems for
parallel processing. Comput. Syst. Engng, 1991, 2(2/3),
135-138.
5. Farhat, C. A simple and efficient automatic FEM domain
decomposer. Comput. Struct., 1988, 28(5), 579-602.
6. Kemighan, B. W. & Lin, S. An efficient heuristic
procedure for partitioning graphs. Bell Syst. Tech. J.,
1970,49, 291-307.
7. Khan, A. I. & Topping, B. H. V. Subdomain generation
for parallel finite element analysis. Comput. Syst. Engng,
1993, 4(4-6), 473-488.
8. Hertz, J., Krogh, A. & Palmer, R. G. Introduction to the
Theory of Neural Computing. Addison-Wesley, Reading,
MA, 1991.
9. Peterson, C. & Anderson, J. R. A mean field learning
algorithm for neural networks. Complex Syst., 1987, 1,
995-1019.
10. Peterson, C. & Anderson, J. R. Neural networks and NPcomplete optimization problems; a performance study on
the graph bisection problem. Complex Syst., 1988, 2(l),
59-89.
11. Khan, AI. & Topping, B. H. V. Parallel adaptive mesh
generation. Comput. Syst. Engng, 1991, 2(l), 75-102.
115