Professional Documents
Culture Documents
(3)
Where Z is the z-axis coordinate value of point to be
computed. Z
max
and Z
min
are z-axis coordinate values of
the highest point and lowest point that present the
maximal value and minimal value of height of points.
According to the Eq.(3), gray value of per pixel can be
calculated. The imageries of scanning scene are showed
in Fig.4.
(a) (b)
(c)
Fig.4. Pseudo-grey imagery of target is generated based on the
height-grey mapping. Images (a), (b) are single target in scanning scene.
Image (c) is two classes of target in one scanning scene.
.TARGET RECOGNITION
The main parts of the framework of target recognition
are illustrated in Fig.5. Imagery of scanning scene has
been gained in Sec.2. In image processing, we adopt the
methods of filtering, thresholding and region labelling.
Median filter is used to remove salt-and-pepper noise.
Weighted average filter is used to smooth the discrete
sampling values of laser points. The imagery after
processing is segmented by heuristic approach, based on
visual inspection of the histogram. The target and
background can be distinguished. Then the targets are
labelled by fixed rectangles. Affine moments of target
region are extracted as features. It has better invariability
than traditional Hu moments in processing the
perspective transformation of images. The improved BP
network algorithm and SVM algorithm are applied to
target classification and recognition.
Fig.5. Main parts in target recognition framework
The original pseudo-grey images are processed. The
results are shown in Fig.6.
(a) (b)
(c)
Fig.6. Original pseudo-grey imageries of target are processed. Images
(a), (b) are single target in scanning scene. Image (c) is two classes of
target in one scanning scene. The targets are labelled with green
rectangles.
For a digital image f(x,y), the moment of order (p+q)
is defined as
( ) ,
p q
pq
x y
m x y f x y dxdy =
(4)
Where p,q=0, 1, 2,
The central moments are defined as
( ) ,
p q
pq
x y
x x y y f x y dxdy u
| | | |
=
| |
\ . \ .
(5)
Where
10
00
m
x
m
= and
01
00
m
y
m
=
The affine moments are defined as [3]
2
20 02 11
1 4
00
A
u u u
u
=
2 2 3 3 2 2
30 03 30 12 21 03 30 12 03 21 21 12
2 10
00
6 4 4 3
A
u u u u u u u u u u u u
u
+ +
=
( ) ( ) ( )
2 2
20 21 03 12 11 30 03 12 21 02 30 12 21
3 7
00
A
u u u u u u u u u u u u u
u
+
=
Image generation
Image processing
Feature extraction
Target recognition
1-614
The Ninth International Conference on Electronic Measurement & Instruments ICEMI2009
Where this set of affine moments
1
A ,
2
A ,
3
A are
invariant to translation, rotation, scale, and affine
transformation.
In the outdoor experiment, we construct a database of
target imagery. As single vehicle rotates per five degrees,
it is scanned by LIDAR sensor. Then two classes of
vehicle are placed together, both of them are scanned in
one scene. The affine moments of target image are
extracted as features for recognition.
(a) (b) (c) (d)
(e) (f) (g) (h)
Fig.7. Four typical poses of targets in images. Image (a) ~ (d) are
imageries of transporter. Images (e) ~ (h) are imageries of racing car.
Target regions have been labelled by green rectangles.
Then the affine moments A
1
,A
2
,A
3
can be calculated.
Traditional Hu moments are also calculated [4]. Here, we
select the four invariants
1
~
4
. Te experimental data
are shown in Table 1 and Table 2.
Table 1. Affine moments and Hu moments of transporter
(a) (b) (c) (d)
A
1
(10
-3
) 7.751 7.612 7.686 7.673
A
2
(10
-8
) 1.340 1.162 1.171 1.326
A
3
(10
-6
) 8.740 8.608 8.783 8.713
1
(10
-1
) 1.852 3.017 3.347 3.420
2
(10
-2
) 0.762 6.215 8.173 8.642
3
(10
-3
) 0.099 2.438 3.153 3.556
4(
10
-3
) 0.002 1.124 1.521 1.741
From the tables, we can know that the affine moment
values A
1
,A
2
,A
3
are very close in the same class and has
large space between different classes. Compared with
traditional Hu moments, the standard deviations of affine
moment are less and therefore, it has excellent
invariability.
Then the affine moments are inputs for BP neural
network and SVM classifier.
BP neural network is a supervised neural network
classifier that learns to classify inputs by given features.
BP neural networks can be used to model complex
relationships between inputs and outputs or to find
patterns in data. In essence, BP neural network can be
viewed as a function that maps input vectors to output
vectors. The knowledge is presented by the
interconnection weight, which is adjusted during the
learning stage using the back-propagation learning
algorithm that uses a gradient search technique to
minimize the mean square error between the actual
output pattern of the network and the desired output
pattern [5, 6].
Table 2. Affine moments and Hu moments of racing car
A
1
(10
-3
) 6.802 6.681 6.747 6.662
A
2
(10
-10
) 3.864 3.565 3.746 3.496
A
3
(10
-6
) 1.324 1.299 1.497 1.308
1
(10
-1
) 3.142 3.486 3.018 3.239
2
(10
-2
) 7.208 9.434 6.429 7.834
3
(10
-4
) 5.731 5.878 5.255 5.257
4(
10
-4
) 3.199 0.596 2.162 2.254
The number of neurons in the input layer and the
output layer is completely determined according to the
requirement of the users. As discussed above, under the
experimental condition, the varied parameters concerned
about the target imagery are affine moments A
1
, A
2
, A
3
.
So the neuron number in the input layer is 3. If the target
is a transporter, we defined the output is (1, 0). If the
target is a racing car, we defined the output is (0, 1). The
number of neurons of output layer is 2. The BP
architecture with one hidden layer is enough for the
majority of applications. Hence, only one hidden layer
was adopted here. Until now, there is not a perfect theory
for instructing choice for the number of neurons in
hidden layers. An empirical equality of number of hidden
layer is shown in Eq.(6). We can finally select a
reasonable value by testing the value that accordance to
the Eq.(6) one by one. The value of L is 6 and the error
between the actual output and desired output has minimal
value.
(1 10) L I O a a = + + s s (6)
Where L is the number of hidden layer, I represents
the number of input layer, O represents the number of
output layer, a represents the constant between 1 and 10.
Among the training algorithms, the
Levenberg-Marquardt (LM) algorithm is the fastest and
least memory consuming one. In this paper, the LM
algorithm was used for training the network. Neural
network toolbox of MATLAB software was used. The
training parameters were set as following: the target
value is 10
-5
, learning ratio is 0.02 and the maximum
training number is 300.
Support vector machine is a very useful technique for
data classification by the principle of structural risk
minimization. SVM-based classifiers are claimed to have
good generalization properties compared to conventional
1-615
The Ninth International Conference on Electronic Measurement & Instruments ICEMI2009
classifiers, because in training the SVM classifier, the
so-called structural misclassification risk is to be
minimized, whereas traditional classifiers are usually
trained so that the empirical risk is minimized[7].
The recognition results are shown in Fig.8.
(a)
(b)
Fig.8 Different types of target are recognized.
We choose 60 images of two classes of target for
testing. The recognition results are given in Table 3.
Table 3. Results of recognition by BP neural network and SVM
BP neural network SVM
Number of correct
recognition
53 58
Number of false recognition 7 2
Total number 60 60
Recognition rate 88.33% 96.67%
The training result for BP neural network is not steady.
And the suitable model for training is difficult to confirm.
The time for computing by BP neural network is more
than SVM. The algorithm of SVM doesnt have the
over-learning phenomena. And recognition rate is higher
than BP neural network. It has advantages in small
samples classification.
.CONCLUSION
In this paper we proposed a sequential approach for
target imagery generation, target classification and
recognition. The processing of data collected by LIDAR
sensor is novel, and it can transform one-dimensional
non-visual original data into two-dimensional visual
imagery. The imagery generation is based on the
principle of height-gray mapping. This method can
effectively distinguish the target from environment.
Targets are classified and recognized by means of
feature-based algorithm. Compared with model-based
algorithm, it has more advantages. The algorithm of
model-based requires storing 3D model of target in
database. These models are often difficult to construct
and not readily available for all types of targets that need
to be recognized. Instead of using a stored model, a
classifier can be constructed to learn a model of each
target type based on a set of observed features from
training imagery. The performance of a classifier in
this feature-based model is dependent on the quality of
the features used.
The affine moments are more excellent than
traditional moments on invariance of perspective
transformation of images. The methods of BP network
and SVM for classification and recognition are
implemented in outdoor experiment. The target
recognition system has been built by the affine moments
and SVM algorithm. In the work described in this paper,
we have succeeded in processing the raw data collected
by LIDAR sensor. In the future we expect to have
non-scanning system to support the imaging speed of
target region. When the imaging speed is available,
real-time performance of recognition system will be
made. If also the system is carried on the aircraft, the
experimental data will be more comprehensive.
REFERENCES
[1] A.Vasile, R.M.Marino. Pose-independent automatic target
detection and recognition using 3-D ladar data[J], Proc. SPIE
5426, 67~83(2004)
[2] R.Shu, Y.Hu, and X.Xue. Scanning modes of an airborne
scanning laser ranging-imager sensor[J], Proc. SPIE 4130,
760~767(2000)
[3] J.Heikkila. Pattern matching with affine moment descriptors[J],
Pattern Recognition, 37(9),1825~1834(2004)
[4] Chen Qing, Petriu Emil, Yang Xiaoli. A comparative study of
fourier descriptors and Hus seven moment invariants for image
recognition[C], Canadian Conference on Electrical and
Computer Engineering, Vol.1, pp: 0103~0106, 02 May,2004.
[5] L.V.Fausett. Fundamentals of neural networks: Architectures,
Algorithms and Applications[M], Prentice Hall press, US. (1994)
[6] S.Haykin. Neural network: A comprehensive foundation(2nd
Edition) [M], Prentice Hall press, US. (1998)
[7] N.Vapnik. The nature of statistical learning theory[M], Springer
press, New York. (1999)
1-616