Professional Documents
Culture Documents
Unit-V
RECOGNITION
Recognition Re + Cognition
Medical diagnosis
Life form analysis
Sonar detection
Radar detection
Image processing
Process control
Information Management systems
Aerial photo interpretation.
Weather prediction
Sensing of life on remote planets.
Behavior analysis
Character recognition
Speech and Speaker recognition etc.
EXAMPLE: DC Machines
Patterns Results
Feature Extraction
Transducer Learning Classification
And
Feature Selection
1.Transformation Method :
Here we reduce the features by considering the linear or nonlinear
combinations of original features. This is also called as aggregation
method.
Eg:- let us assume originally we have four features f1,f2,f3,f4.
One method of selecting two features is
f5 = f1 + f2
f6 = f3 + f4.
2.Subsetting or filtering Method:
Here we select a subset of the original features.
Eg:- Original features are f1,f2,f3,f4.
We can select a subset like
f5 = f1 and f6 = f3.
Learning : It is a process of determining useful parameters which are required
for classifying the patterns efficiently.
Classifying: Here the patterns are assigned to different classes using a
suitable classification method.
PRINCIPAL COMPONENT
ANALYSIS (PCA)
Motive
Find bases which has high variance in data
Encode data with small number of bases with low MSE
Assume that
1
E[ x ] 0 T
ax qq x T T
|| q || (q q) 2 1
2 E[a 2 ] E[a ]2 E[ a 2 ]
E[(qT x)(xT q)] qT E[xxT ]q qT Rq
Find qs maximizing this!!
Principal component q can be obtained
by Eigenvector decomposition such as SVD!
Rq q
Neelam Rawat, AI UNIT - V 18
Dimensionality Reduction
(1/2)
You do lose some information, but if the eigenvalues are small, you
dont lose much
n dimensions in original data
calculate n eigenvectors and eigenvalues
choose only the first p eigenvectors, based on their
eigenvalues
final data set has only p dimensions
Neelam Rawat, AI UNIT - V 19
Dimensionality Reduction
(2/2)
Variance
Dimensionality
Original
q=16 q=32 q=64 q=100 Image
c ni
Within-class scatter matrix Sw (Y
i 1 j 1
j M i )(Y j M i )T
c
Between-class scatter matrix Sb (M
i 1
i M )( M i M )T
projection matrix
y U x T
PCA LDA
Let data D be (X1, y1), , (X|D|, y|D|), where Xi is the set of training
tuples associated with the class labels yi
There are infinite lines (hyperplanes) separating the two classes but
we want to find the best one (the one that minimizes classification
error on unseen data)
SVM searches for the hyperplane with the largest margin, i.e.,
maximum marginal hyperplane (MMH)
Neelam Rawat, AI UNIT - V 35
SVM Linearly Separable
_
_
_ _ .
+
_ . +
xq + . . .
Neelam Rawat, AI UNIT - V . 38
Discussion on the k-NN Algorithm