Professional Documents
Culture Documents
Introduction
478
Z. Wang et al.
curve. It will be demonstrated that the representation can be scale, rotation and
translation invariant after proper normalizations. Particularly, a thinning-based
method is adopted to locate the start point(s) for reducing the computation time
in image retrieval. In order to further reduce the retrieval time, we then propose
a two-step approach which uses both the centroid-contour distance curve and
the eccentricity of the leaf object for shape-based leaf image retrieval.
In Section 2, we define the centroi-contour distance curve and explain its
invariant properties. A similarity measure with the distance curve is also discussed in the section. A leaf image retrieval scheme based on the eccentricity
and centroid-contour distance curve is presented in Section 3. Experimental results and discussions are given in Section 4. Finally, concluding remarks are
drawn in Section 5.
2
2.1
Tracing the leaf contour can be considered as circling around its centroid. The
trace path from fixed start point represents a shape contour uniquely, that is to
say, a contour point sequence corresponds to a shape uniquely if the start point
is fixed. This is the basic idea for chain code representation of shape. As Figure
1 shows, a point P on the contour is determined by the centroid C, the distance
R between the point P and the centroid C, and the angle . In fact, it is not
necessary that the object centroid has to be fixed in a coordinate system since the
change of object centroid only leads to the translation of the object. The shape
reconstruction can also be independent of the angle given the fixed start point.
Therefore, for the same start point, the object contour can be resonstructed with
the centroid-contour distance curve. The distance between the object centroid
and a contour point is termed as the centroid-contour distance. The contour can
be represented by one dimensional curve, the centroid-contour distance curve.
2.2
Generally, the properties of scale, translation and rotation invariance are expected for the shape feature based image retrieval. After some normalization, we can
achieve these invariant properties with the centroid-contour distance curve.
Translation Invariant Property
The contour shape of an object is fully determined by the centroid-contour distance function and is nothing to do with the coordinates of the centroid position.
Therefore, the distance curve is translation invariant. This is elaborated as follows. As shown in Figure 1,
p
|CP | = (xC xP )2 + (yC yP )2
(1)
P1
R
P
C1
R
a
C
479
y
x
where (xC , yC ) and (xP , yP ) are the coordinates of points C and P respectively.
And the object centroid (xC , yC ) is defined as follows.
Z Z
m10
m01
mpq =
xp y q dxdy, xC =
, yC =
(2)
m00
m00
R
If the object is traslated by increasements x and y along x-axis and yaxis, repectively. The point P on the object contour is moved to point P1 with
its coordinates being (xP + x, yP + y). According to Equation 2,
RR
RR
(x + x)dxdy
m110
1 xdxdy
R
RR R
xC1 = 1 = R R
=
m00
dxdy
dxdy
R1
R
RR
RR
xdxdy
(x)dxdy
(3)
= R RR
+ RRR
= xC + x
dxdy
dxdy
R
R
where R1 is the region of the new positioned object.
Similarly, yC1 = yC + y, that is, the new centroid point C1 of the object is
(xC + x, yC + y). Obviously,
p
|C1 P1 | = (xC1 xP1 )2 + (yC1 yP1 )2
p
= (xC xP )2 + (yC yP )2
= |CP |
(4)
The above equations show that the translation invariant property of the centroidcontour distance curve.
Scale Invariant Property
Let us consider a circle first. The centroid-contour distance of a circle object is
a constant. For a larger circle, its distance is a larger constant and there are
480
Z. Wang et al.
Y
C
A
D
B
O
(a)
(c)
481
(b)
(d)
(e)
Fig. 3. Scale invariant property of center distance curve. (a) Original image; (b) Halfsize image; (c) Centroid-contour distance curve for original image; (d) Centroid-contour
distance curve for half size image; (e) Sampled center distance curve of original image.
Pm
Pm1
P1
Pm
P1
a
O
a
O
P2
482
Z. Wang et al.
(a)
(b)
(c)
(a) Skeleton
(b) Contour
2.3
483
Similarity Measurement
Two-step retrieval
Retrieval
results
Retrieve top m
Eccentricity
candidates in terms
of eccentricity
Centroid-contour
distance curve
(x xC )p (y yC )q dxdy
(6)
where R is the whole region of an object and A is its area. and (xC , yC ) is the
object centroid which is the same as that used in computing the centroid-contour
484
Z. Wang et al.
w1 De (I, J) + w2 Dc (I, J)
w1 + w2
(8)
where I and J denotes two leaf images, and De and Dc are the distance measures
with two features. w1 and w2 are used to weigh the relative importance of two
features,which are determined by simulation tests.
Figure 7 shows the flow chart of our approach. In terms of the eccentricity measure, the top scored M images are retrieved from the image database.
From these M images, the top scored N (N < M ) images are selected based on
Ds (I, J) defined in Equation 8. In order to reduce search time furtherly, images in the database can be indexed with the eccentricity values of the images.
When a query is submitted, the leaf image with the closest eccentricity value
can be found with a fast search algorithm such as the half search algorithm.
The top M leaf images with close eccentricity values can be easily found nearby. Since eccentricity is one dimensional feature, a multi-dimensional k-nearest
search algorithm, such as the k-d tree approach [10], is not necessary.
In our experiments, two data set are used to test our approach. Data set 1
contains 135 320 240 color leaf images that seldomly come from the same
plant. Data set 2 containing 233 color leaf images in arbitary sizes with about 10
samples collected from each plant. From Table 1 we can see that the thinningbased start point(s) locating method has reduced the shift-and-matching times
greatly.
When a query request is submitted, the two-setp retrieval approach is performed. In the experiments, the top 30 closest images will be returned in the
first step.
Figure 8 shows retrieval results for a query performed on the data set 1.The
retrieved images are circularly except that the boundary of image Figure 8(f) are
round. This result indicates that the features pay more attention to the global
shape information. Another retrieval example performed on the data set 2 is
485
Data Set 1
Data Set 2
(a)
(b) 0.2128
(c) 0.2191
(d) 0.2282
(e) 0.2393
(f) 0.2592
(g) 0.2639
Fig. 8. A retrieval example for a sawtooth contour shape in the data set 1. (a) Query
image; (b)-(g) Top 6 retrieved images with their Ds values.
(a)
(b) 0.0494
(c) 0.0630
(d) 0.0661
(e) 0.0692
(f) 0.0740
(g) 0.0758
Fig. 9. A retrieval example for leaf image with stiff corners with the data set 2. (a)
Query image; (b)-(g) Top 6 retrieved images with their Ds values.
486
Z. Wang et al.
shown in Figure 9. We can find the top six images are with the same plant. In
these experiments, we set w1 = 0.4, w2 = 0.6 emperically.
These examples show that our approach can achieve retrieval results that are
similar to the results from human visual perception. Table 2 shows the retrieval
time for the above two examples when the exhaust search and our two-step search
scheme are used. Our experiments are carried out on a Pentium 333 HZ PC. It
is found that the two-step retrieval scheme can reduce the retrieval time significantly. The experimental results show that our approach is computationally
more efficient.
Table 2. Comparison of retrieval time.
Exhaust search (Sec.) Two-step search (Sec.)
Example 1
13.32
2.82
Example 2
13.14
3.73
Conclusion
Acknowledgment
The work described in this paper was substantially supported by a grant from
the Hong Kong Polytechnic University (Project No. G-V780).
References
1. A. K. Jain. Fundamentals of Digital Image Processing. Prentice Hall, London, UK,
1989.
487