Professional Documents
Culture Documents
Abstract
Kernel-based nonlinear feature extraction and
classification algorithms are a popular new research
direction in machine learning and widely used in many
fields. Firstly, we will give an overview of kernel
Fisher discriminant analysis and Support Vector
Machine, and then the description of multi-class
classification method applied for them. The
performance of these two classification method is
analyzed on CASIA II database.
1. Introduction
To meet the increasing security requirement of
current commercial society, personal identification is
becoming more and more important. Some traditional
methods are usually not reliable. Therefore a new
method for personal identification named biometrics
has been attracting more and more attention. Thus, iris
patterns, due to the complexity of two underlying
processes, are believed to be unique.
Flom and Safir first proposed the concept of
automated iris recognition in 1987 [1]. Since then, iris
recognition has been receiving many researchers
attention [2-4]. The first complete iris based
recognition system was designed and patented by
J.Daugman [2]. He utilized Gabor filter with different
frequency and phase to denote the structure
information of iris and measure the resulting code with
Hamming distance. The algorithm of 1-D wavelet
transform at various levels on iris image is proposed
by Boles and Boashash [3]. He made use of the
projection of iris texture in low-dimension space as
characteristics and using Euclidean distance as
matching rule. V.Dorairaj and N.A.Schmid use PCA
and ICA for encoding and use Euclidean distance and
Hamming distance as measures [4].
In this paper, we propose a novel method for iris
recognition, where wavelet transform and PCA is used
for feature extraction, while kernel fisher discriminant
analysis or SVM is used for classification. In order to
illustrate the potential of SVM and KFD for iris
recognition, we give some experimental results on the
proposed method and comparison between existing
methods and proposed method.
____________________________________
0-7803-9737-1/06/$20.00 2006 IEEE
2. Kernel-based Classifier
2.1 Kernel -idea
The idea of Kernel, an efficient nonlinear mapping
method, is firstly used in Support Vector Machine.
Through nonlinearly mapping using kernel trick, the
original space can be transformed into arbitrary higher
dimensional feature space. According to covers
theorem on the reparability of patterns, nonlinearly
separable patterns in the original space turn to be
linearly separable in a higher dimensional space. The
nonlinear mapping between the input space and the
feature space, with a possibly prohibitive
computational cost, is never implemented explicitly by
Kernel-idea [5, 7].
X = [ x0 , x1 ,LL x M 1 ] C N 1 be the
training samples, and be the nonlinear mapping
Let
< ( x i ), ( x j ) >= k ( x i , x j ) .
(1)
(( x + y ) + ) d
exp{
x y
(2)
2
2 2
Sigmoid kernels:
tanh( ( x y ) + )
where, d N , > 0, > 0.
(3)
(4)
XY
= {( ( x1 ), y1 ), L , ( ( x n ), y n )} y {1,2}
wT S w
J F (W ) = T B ,
w SW w
(5)
S B = (u1 u 2 )(u1 u 2 ) T ,
u y =
1
ly
(6)
y 2 closest to H locate at H 2 : wx + b = 1 + i .
The Optimal Hyperplane can be obtained by minimize
the following quadratic programming task:
n
1
( w, b) = arg min( w w + c i p )
2
i =1
c > 0, i > 0
where,
L=
SW = S1 + S 2
(7)
ly
S y = ( ( xi ) u y )( ( xi ) u y ) T , y {1,2} .
i =1
i ( xi ) , so the class
~
T K (J 1)K T
(8)
T K (I J )K T
~
where, K is the kernel matrix, and 1 is k k matrix
with element 1 / k , and J is
1 / k y , if < y = yi = y j >
J ij =
y {1,2} (9)
0
,
otherwise
Q( ) = i
with restrict i
* which maximize
1
i j y i y j ( xi x j )
2 i, j
(13)
i 0, i [1 i y i ( wxi b)] = 0,
i 0, i i = 0
f ( x) = sgn( i* y i ( xi x) + b * )
(14)
i =1
where,
x is a new sample.
> 0 , corresponding x are support vector.
For i
(15)
i =1
f ( x) = sgn( i* y i k ( xi x) + b * )
Through
(12)
i =1
J F (W ) =
1
|| w || 2 +C i +
2
i
i [1 i yi (wxi b)] i i
correctly.
function,
are slack variables used to solve the problem of nonseparable data [9]. Through bringing into lagrange rule,
the optimal problem turn into maximize the following
function:
,l y
i =1
(11)
w ,b
( x ), y {1,2} ,
Because
y1 closest to H locate
the
Hyperplane
y i ( wxi + b) 1 i ,
(10)
as
jth class
x will
be separated into ith class if only ith row of L have
most number of '1' . For a new sample x , we use the
set to 1; otherwise, L ji will be set to 1. Finally,
k ( k 1) / 2
discriminant
functions
to
estimate,
(17)
x k , k = 1,2 L n ,
= 0.
k =1
C=
1 n
xi xiT ,
n i =1
(18)
Methods
Daugmans
Zhus
PCA
ICA
KFD
SVM
4. Conclusion
In this paper, a new method for iris recognition is
proposed, where wavelet-transform and PCA is used
for feature extraction and dimension reduction, while
KFD or SVM is used for classification. Experiments
illustrate that on the basis of wavelet transform and
PCA, KFD and SVM can improve the performance
and outperform the existing method for iris recognition,
where SVM with Gaussian kernel function and OAO
multi-class method performs best.
Acknowledgment
This work is supported by the National Natural
Science Foundation of China (No.60472046).
References
[1]
Multiclass
Method
OAA
OAO
Different
kernel
function
Linear
Poly (d=3)
Gaussian
Sigmoid
Linear
Poly (d=3)
Gaussian
Sigmoid
FAR
0.08%
0%
0%
0.08%
0.08%
0.50%
0%
0.67%
FRR
2.83%
6.00%
2.08%
3.08%
0.67%
4.50%
0.42%
2.08%
Total
Error
Rate
2.92%
6.00%
2.08%
3.16%
0.75%
5.00%
0.42%
2.75%
[2]
[3]
[4]
[5]
[6]
Error Rate
4.08% (EER)
19.86%(EER)
3.84%(EER)
5.21%(EER)
0.42%(FAR) 0.58%(FRR)
0% (FAR) 0.42%(FRR)
[7]
[8]