Professional Documents
Culture Documents
1. Introduction
Linear discriminant analysis (LDA) is a well-known
scheme for feature extraction and dimensionality reduction
of labeled data in a vector space. Therefore, if each datum
is represented as matrix, then it must be transformed into
a vector. However, this transformation from a matrix to a
vector might cause the so-called singularity problem or the
small sample size (S3) problem in which the scatter matrices become singular and thus the traditional LDA fails to
work well. To address this problem, several extensions of
LDA have been presented recently. Li and Yuan [2] and
Xiong et al. [6] referred to 2DPCA [8] and presented twodimensional linear discriminant analyses based on image
between-class scatter matrix and image within-class scatter
matrix. Their methods reduce the dimensionality of image
matrices without converting the image matrices into vectors. Thus the singularity problem resulting from the highdimensionality of vectors is artfully avoided. Liu et al. [3]
also presented a similar method to theirs more than ten years
ago. Yang et al. [7] and Song et al. [5] presented their
variants. However, their methods reduce only the number
of columns in each matrix and the number of rows is unchanged. Yang et al. [9] presented a two-stage algorithm
and named it two-dimensional linear discriminant analysis
(2DLDA), which reduces the number of columns first and
2. An Overview of 2DLDA
In this section, we give a brief overview of 2DLDA presented by Ye et al. [10].
Let
, for , be the images in
a dataset, clustered into classes , where has
images. Let
be the mean of the
0.93
0.92
0.91
f(L,R)
th class, , and
0.9
0.89
0.88
0.87
5
10
15
number of iterations
20
(1)
(2)
where denotes the matrix trace. The optimal transforma . Due
tions and
would maximize
to the difficulty of computing the optimal and
simultaneously, Ye et al. [10] derived the following iterative algorithm. Initially, we fix
. Then we can compute
the optimal . Next, we fix the and compute the optimal
Step 2:
Compute
and
.
Step 3: Compute the first eigenvectors
of
and form .
Step 4:
Compute
and
.
Step 5: Compute the first eigenvectors
of
and form
.
Step 6: Increase by 1. If then go to Step 2,
otherwise output and
, where
is the maximal number of iterations.
However, this iterative algorithm does not necessarily increase
monotonically. An example with the ORL
face image database [4] is shown in Fig. 1, where the horizontal axis is the number of iterations and the vertical axis
in this example. This
is
. We set
non-monotonicity corroborates the perturbation in the accuracy curve of 2DLDA reported in [10]. Due to the nonmonotonicity, it is hard to determine appropriate criteria for
stopping the iteration. Therefore, Ye et al. [10] adopted a
simple algorithm that outputs the result of the first iteration.
Figure 1. Variations in
.
3. Non-Iterative 2DLDA
In this section, we propose two non-iterative algorithms
for 2DLDA, i.e., the selective and parallel algorithms.
The first iteration of 2DLDA computes and
in turn
with the initialization
. Alternatively, we can
consider another algorithm that computes
and in turn
with the initialization . This alternative algorithm is equivalent to another 2DLDA presented by Yang et
al. [9]. In this subsection, we propose an algorithm unifying
them. That is, we select and
which give larger
.
The selective algorithm is as follows:
Selective Algorithm
and compute and
in
Step 1: Initialize
turn. Let and
be the computed and
.
and compute
and in
Step 2: Initialize
turn. Let and
be the computed and
.
Step 3: If
then output
and
, otherwise output and
.
In this subsection, we propose another non-iterative algorithm, which computes and
independently. Let us
define the row-row within-class and between-class scatter
matrices as follows:
(3)
(4)
(5)
where is the identity matrix of size . Let
be the eigen-decomposition of , where is a diagonal
matrix whose diagonal elements are eigenvalues of and
is an orthonormal matrix whose columns are the corre into
sponding eigenvectors. Substitution of
(5) gives
(6)
Let
be the first
eigenvectors of
. Then the optimal solution of
. Consequently we can
(6) is expressed as
.
calculate the optimal solution of (5) as
(7)
(9) gives
(10)
Let be the first eigenvectors of
. Then the optimal solution of
. Consequently we
(10) is expressed as
(11)
where
and denotes the Frobenius norm.
(8)
(9)
where is the identity matrix of size . Let
be the eigen-decomposition of , where is a diagonal
matrix whose diagonal elements are eigenvalues of and
is an orthonormal matrix whose columns are the corre
into
sponding eigenvectors. Substitution of
4. Experimental Results
.
can calculate the optimal solution of (9) as
In this section, we experimentally evaluate the performance of iterative and non-iterative 2DLDA on the ORL
face image database. The ORL database [4] contains face
images of 40 persons. Each person has 10 different images.
That is, the total number of the images is 400. The size of
and . Some
each image is , i.e.,
example images are shown in Fig. 2. We select five images
per person for training, and the remaining five images for
testing. So the sizes of training and testing data are both
200, i.e., . We make and the same size for
simplifying our experiments. Figure 3 shows the result of
face recognition. The horizontal axis denotes and the vertical axis denotes the recognition rate. We fixed the number
of iterations of the iterative 2DLDA for 10. The recognition rate of the selective algorithm denoted by broken line
is higher than that of the iterative 2DLDA denoted by solid
line on the whole. The recognition rate of the parallel algorithm denoted by dotted line is a little lower than that of the
iterative 2DLDA in the wide range of .
We performed our experiments using Matlab on a Pentium IV 3.4GHz machine with 2GB RAM. Figure 4 shows
the CPU time for training. This figure shows that the selective algorithm denoted by broken line is computationally
3
2.5
2
2DLDA
selective
parallel
1.5
1
0.5
0
2
8 10 12 14 16
p
recognition rate
1
0.9
0.8
0.7
2DLDA
selective
parallel
0.6
0.5
2
10 12 14 16
5. Conclusion
In this paper, we presented two non-iterative algorithms
for 2DLDA, i.e., the selective and parallel algorithms. It is
experimentally shown that our algorithms achieve competitive recognition rates with the iterative 2DLDA, while the
CPU times of our algorithms are much shorter than that of
the iterative 2DLDA.
2DLDA can be regarded as a sort of bilinear discriminant
analysis. Extensions of 2DLDA to multilinear discriminant
analyses are future works.
References
[1] C. Bauckhage and J. K. Tsotsos. Separable linear discriminant classification. DAGM 2005, LNCS, 3663:318325,
2005.
[2] M. Li and B. Yuan. 2D-LDA: A novel statistical linear discriminant analysis for image matrix. Pattern Recognition
Letters, 26(5):527532, 2005.
[3] K. Liu, Y. Cheng, and J. Yang. Algebraic feature extraction for image recognition based on an optimal discriminant
criterion. Pattern Recognition, 26(6):903911, 1993.
[4] F. Samaria and A. Harter. Parameterisation of a stochastic
model for human face identification. Proceedings of 2nd
IEEE Workshop on Applications of Computer Vision, 1994.
[5] F. Song, S. Liu, and J. Yang. Orthogonalized Fisher discriminant. Pattern Recognition, 38(2):311313, 2005.
[6] H. Xiong, M. Swamy, and M. Ahmad. Two-dimensional
FLD for face recognition. Pattern Recognition, 38(7):1121
1124, 2005.
[7] J. Yang, J. Yang, A. F. Frangi, and D. Zhang. Uncorrelated
projection discriminant analysis and its application to face
image feature extraction. International Journal of Pattern
Recognition and Artificial Intelligence, 17(8):13251347,
2003.
[8] J. Yang, D. Zhang, A. F. Frangi, and J. Yang. Twodimensional PCA: A new approach to appearance-based
face representation and recognition. IEEE Trans. Pattern
Anal. and Machine Intell., 26(1):131137, 2004.
[9] J. Yang, D. Zhang, X. Yong, and J. Yang. Two-dimensional
linear discriminant transform for face recognition. Pattern
Recognition, 38(7):11251129, 2005.
[10] J. Ye, R. Janardan, and Q. Li. Two-dimensional linear discriminant analysis. Advances in Neural Information Processing Systems (NIPS 2004), 17:15691576, 2004.