Professional Documents
Culture Documents
in
data
the
algorithms.
preprocessing
performance
Many
of
techniques
plays
the
have
vital
machine
been
role
for
learning
proposed
by
in accuracy, it
for
building
the
learning
model
in
the time
machine
performance of these
Cfs,
Consistency
and
Machine learning;Classifier.
I.
INTRODUCTION
125
;=1
(1)
126
II.
RELATED WORK
Cjs
kr
k+k(k-l)rjJ
cf
(2)
2.0x
(3)
B. Consistency
In consistency-based subset elevation, many approaches
use class consistency as an evaluation metric in order to
select the attribute subset [40], [41]. These methods look
for combinations of attributes whose values divide the data
into subsets containing a strong single class majority.
several approaches to attribute subset
selection use class consistency as an evaluation metric [40],
[41]. These methods look for combinations of attributes
whose values divide the data into subsets containing a
strong single class majority. Usually the search is biased in
favour of small feature subsets with high-class consistency.
Our consistency-based subset evaluator uses the work of
[41] consistency metric.
J
Consistencys
IID,I-IM,I
1--"'="'----
(4)
Filter
III.
PROPOSED WORK
127
IV.
S.No.
Dataset
Contact Lenses
2
3
4
5
6
7
Diabetes
Glass
Ionosphere
Iris
Labor
Soybean
8
9
10
Super Market
Vote
Weather
3S
30
2S
20
IS
Instances
a Cis
10
Features
Classes
24
768
214
351
150
57
683
9
10
35
5
17
36
2
7
2
3
2
19
4627
435
14
217
17
5
2
2
2
S
0
II
--.11
...
c,t." ,,',,'1 e" '$:-" '"
o"Q
..."t (J-li
,o
<..0r.:-
S.No.
115
'i- 77.5
,
77
Cfs
Collsistellcy
filtered
Contact Lenses
Diabetes
3
4
Glass
Ionosphere
8
14
7
7
5
14
5
6
Iris
Labor
2
7
2
4
2
4
7
8
Soybean
Super Market
22
13
32
7
15
Vote
Weather
4
2
10
2
1
1
9
10
o\ ::o,,'b(;o t""
'" ,,0"
II ..
II Consisttncv
WFiltered
o-$- t\
,,'&
'l
79
.... Y
n--
795
TABLE fl.
11
76.5
"
755
75
'45
,
'"
Consisttncy filt!r!d I
Cfs
"""'""
C4.S(J48)
NO
'ii,,,.,
'"
Coo""'""
'81
" .,,'
V.
CONCLUSION
128
2012 IEEE
TABLE lll.
S.No.
I
2
3
4
5
6
7
8
9
10
Dataset
Contact Lenses
Diabetes
Glass
Ionosphere
Iris
Labor
Soybean
Super Market
Vote
Weather
Avera2e
Accuracy of NB of Redacted
Feature Subsets
Accuracy of C4.5(J48) of
Redacted Feature Subsets
COllsistellcy
Filter
Cfs
COllsistellcy
70.83
8333
70.83
66.66
62.50
66.66
74.86
68.69
74.86
70.09
74.60
65.88
68.35
71.02
68.35
70.09
70.18
71.02
Cfs
COllsistellcy
Filter
Cfs
70.83
70.83
70.83
77.47
47.66
77.47
44.39
76.43
44.39
Filter
92.02
87.17
92.02
90.59
87.46
90.59
88.88
87.74
88.88
96.00
91.22
87.11
96.00
87.71
81.69
96.00
87.71
83.30
96.00
77.19
85.65
96.00
82.45
83.74
96.00
80.70
82.86
96.66
84.21
83.89
96.66
87.71
76.57
96.66
87.71
79.94
63.71
96.09
57.14
63.71
92.41
57.14
63.71
95.63
50.00
63.71
96.09
42.85
63.71
96.32
42.85
63.71
95.63
50.00
57.20
94.02
78.57
43.00
93.33
78.57
53.07
91.72
64.28
77.92
75.85
76.00
76.649
78.08
77.08
78.95
76.45
77.018
REFERENCES
[1]
[11] Md. Monrul Kabir,Md. Shahj ahan,Kazuyuki Murase "A new hybrid
ant colony optimization algorithm for feature selection" Expert
systems with application 39 3747-3763 (2012)
[2]
[3]
[4]
[5]
[6]
[7]
Xuechuan Wang
Kuldip K. Paliwal "Feature extraction and
dimensionality reduction algorithms and their applications in vowel
recognition" Pattern Recognition 36 (2003) 2429 - 2439
[8]
[9]
[10] Oh,l, Lee,.J & Moon B (2004)" Hybrid genetic algorithms for
feature selection" lEE transaction on Pattern Analysis and Machine
Intelligence 26(11). 1424-1437
129
2012 IEEE
[28] Dash, M., & Liu, H. "Feature selection for classification. "
Intelligent Data Analysis (1) 131-156 (1997)
[29] Huang, J., Cai,Y., & Su,X. A hybrid genetic algorithm for feature
selction wrapper based on mutual information. Pattern Recognition
letters, 28,1825-1844 (20 0 7)
[30 ] Guan,1 S., Liu, J., & Qi, Y
"An incremental approach to
contribution-based feature selection" Joumal of Intelligence System
13(1) (20 0 4)
[3 I] Peng, H., Long, F., & ding. C "Overfiting in making comparisions
between variable selection methods. " Journal of Machine Learning
Research, 3,137-1382.(200 3)
[32] Gasca, E., Sanchez, J.S., & Alonso R. "Elimination redundancy and
irrelelevance using a new MLP-based feature selection method. "
Pattern Recognition, 39 , 313-315 (20 06)
[33] Hsu, C, Huang. H., & Schuschel, D. "The ANNIGMA-wrapper
approach to fast feature seletion for neural nets". IEEE
Transactions on system, Man, and Cybemetics- Part B: Cybernetic,
32(2) 20 7-212(20 02)
[34] Caruana, R., &
Freitage, D "Greedy attribute selection". In
proceedings of 11"' intemational conference of machine learing
USA: Morgan Kaufman (19 9 4)
[35] Lai, C, Reinders, M.J.T., & Wessels, "L Random subspace method
for multivariate feature selection". Pattern Recognition Letters , 27,
1067-1076 (20 06)
[36] Straceezzi, D.
elimination. "
J., &
Utgoff, P.
E.
"Randomixed
variable
[37] Liu, H., & Tu, L. Toward integrating feature selection algorithems
for classification and clustering . IEEE Transactions on Knowledge
and Datat Engineering 17(4),491-502.(2004)
[38] M. A. Hall, "Correlation-based feature selection for mL/chine
learning," Ph.D. thesis,
Department of Computer Science,
University of Waikato, Hamilton, New Zealand, 199 8.
130