Professional Documents
Culture Documents
I. I NTRODUCTION
Predicting the future rather than describing the data at
hand is the primary goal of learning but many statistical
learning problems involve some form of dimension reduction
to achieve this goal, which may involve feature selection
with the aim of finding similarities of the original set of
variables [2] and a new set of data to appropriately predict
future behaviors of variables. Finding an accurate model, faces
challenges of having a much smaller sample data in a much
higher dimensional space that results in strong reaction if
the model is too highly-parameterized which result in overfitting of the data and fails to learn the function features
[1] [8] . However model with too few parameters may not
even describe the training data adequately and also provide
similar performance. With regularization we can balance a
reliable way to create a model that evaluates the method for
complexity matching for each class depending n the dimension
of the data been processed. Therefore, kernels will provide a
computationally feasible and flexible method for implementing
such algorithm. Reproducing Kernel Hilbert Spaces (RKHS)
introduces a particular useful set of hypothesis associated with
a particular kernel, and using this in deriving the general
solution of Tikhonov regularization in RKHS will help us
implement and solve the above statistical complications [6].
A. Kernels
Considering RKHS theory, let V Rn be endowed with
an inner product. The configuration involves three aspects: the
vector V , the orientation of V in Rn , and the inner product
on V all with a basis v1 , ..., vr V Rn for V and the
The work in this paper was aided by several online materials including
some of the proofs which can be provided on request
(1)
n
X
K(xi , x) ,
(2)
i=1
b=
1
y wT X 1 = y wT x
n + 0
(11)
y=
1
y1,
n + 0
(12)
x=
1
X1
n + 0
(13)
where
and
i,j=1
(4)
for all x, x0 X.
As discussed in Section 1, regularization is introduces to
prevent over-fitting by introducing additional constraints on
the model as a form of complexity penalty. Considering the
classifier given in the homework:
yj = wT (xj ) + b.
(5)
y = wT
1 T
= T T + I
y
(10)
= k (K + I)
(15)
1
yT
(16)
(17)
(18)
1
y wT 1 = y wT
n + 0
(19)
b=
(20)
j=1
Fig. 6: The double Fibonacci spiral for N = 400 data using the reproducing
kernel Hilbert space.
Fig. 8: The double Fibonacci spiral for N = 400 data using the SVM with
Gaussian kernel.
SVM
N3
0.50
C= 5 and = 0.1
RKHS
N2
0.5012
= 0.01 and = 0.1