Professional Documents
Culture Documents
Analysis
T.S. Yo
References
Outline
● Introduction
● Learn the distance metric from data
● The size of K
● Procedure of NCA
● Experiments
● Discussions
Introduction (1/2)
● KNN
– Simple and effective
– Nonlinear decision surface
– Non-parametric
– Quality improved with more data
– Only one parameter, K -> easy for tuning
Introduction (2/2)
● Drawbacks of KNN
– Computationally expensive: search through the
whole training data in the test time
– How to define the “distance” properly?
Softmax Function
1
0.9
0.7
0.6
exp(-X)
0.5
0.4
0.3
0.2
0.1
0
X
Learn the Distance from Data (4/5)
● How do we define dij ?
● Limit the distance measure within Mahalanobis
(quadratic) distance.
n2d is a mixture of two bivariate normal distributions with different means and
covariance matrices. ring consists of 2-d concentric rings and 8 dimensions of
uniform random noise.
Experiments – Results (1/4)
● Results of
Goldberger
et al.
(rank 2
transformation)
Discussions (5/8)
● Results of experiments suggest that with the
learned distance metric by NCA algorithm, KNN
classification can be improved.
– Laplacianfaces(LAP)
● Emphasizes more on dimension reduction