You are on page 1of 15

International Journal of Engineering and Techniques - Volume 3 Issue 5, Sep - Oct 2017


Improved Genetic Algorithm Based Feature Selection

Strategy Based Five Layered Artificial Neural Network
Classifier (Iga – Flann)
M. Praveena 1, Dr.V.Jaiganesh2
Assistant Professor, Dept. of Computer Science, Dr.SNS Rajalakshmi College of Arts and Science,Coimbatore,
Tamil Nadu, India.
Professor, Dept. of PG & Research, Dr. NGP Arts and Science College, Coimbatore,Tamil Nadu, India.

Data classification is one of the investment research areas in the field of data mining. Machine learning
algorithms such as naive bayes, neural network, and support vector machine are most regularly used for performing
the classification task. Supervised learning is one of its kinds where the datasets consist of class labels and the
machine learning classifier are trained first using that. It is to be noted that feature selection plays a vital role in
developing the classification accuracy of the supervised machine learning classifiers. This research work aims in
proposing an improved genetic algorithm based feature selection planning based five layered artificial neural
network classifier. Around 20 datasets are collect from the UCI repository. Implementations are carried out using
MATLAB tool. Performance metrics such as prediction efficiency and time taken for prediction are taken into
account to conduct the performance evaluation of the expected classifier. Simulation results portrays that the
proposed IGA-FLANN classifier outperforms the existing classifiers.

Keywords — Machine learning, boosting, neural network, genetic algorithm, feature selection, data mining,

I. INTRODUCTION fangled data. ML algorithms are broadly

classified into three categories namely
Data mining is one among the thrust supervised learning, unsupervised learning
research areas in the field of computer and reinforcement learning and is shown
science. Yet there is a knowledge data in Fig.1. The progression of machine
discovery process helps the data mining to learning is comparable to that of data
abstract hidden information from the mining. Both data mining and machine
dataset there is a huge scope of machine learning consider or explore from end to
learning algorithms. Particularly end data to seem for patterns. On the other
supervised machine learning algorithms hand, in preference to extracting data for
gain major importance in data mining human knowledge as is the case in data
research. Machine learning shortly mining applications; machine learning
describe as ML is a kind of artificial makes use of the data to discover patterns
intelligence (AI) which makes available in data and fine-tune program actions
computers with the efficiency to be trained therefore.
without being overtly programmed. ML
learning spotlights on the development of
computer programs which is capable
enough to modify when disclosed to new-

ISSN: 2395-1303 Page 199

International Journal of Engineering and Techniques - Volume 3 Issue 5, Sep - Oct 2017


2.1. Recent Works on Decision Trees

Lertworaprachaya et al., 2014 proposed a

new model for compose decision trees
using interval-valued fuzzy membership
values. Most existing fuzzy decision trees
do not examine the uncertainty associated
with their membership values; however,
correct values of fuzzy membership values
are not always possible. Because of that,
Fig.1. Machine Learning and its Types
the authors represented fuzzy membership
values as distance to model uncertainty
Supervised machine learning is the and employ the look-ahead based fuzzy
mission of assume a meaning from decision tree induction method to construct
labelled training data which has a set of decision trees. The authors also inspected
training examples. As far as supervised the significance of different
learning is concerned, every example is a neighbourhood values and define a new
brace containing an input object (which is parameter unkind to specific data sets
usually a vector quantity) and a using fuzzy sets. Some examples are
compulsory output value (may also be provided to establish the effectiveness of
referred as supervisory signal). A their approach.
supervised learning algorithm at first
performs the analysis task from the Bahnsen et al. 2015 proposed an example-
training data and composes a contingent dependent cost-sensitive decision tree
function, in order to map new examples. algorithm, by incorporating the different
An optimal setting apparently facilitates example-dependent costs into a new cost-
the algorithm to exactly courage the class based corruption measure and new cost-
labels for concealed instances and the based pruning criteria. Subsequently, using
same needs the supervised learning three different databases, from three real-
algorithm to reduce from the training data world applications namely credit card
to concealed situations in a "rational" fraud detection, credit scoring and direct
manner. The supervised methods are marketing, the authors calculate their
perhaps used in various application areas proposed method. Their results showed
that include marketing, finance, that their proposed algorithm is the best
manufacturing, testing, stock market performing method for all databases.
prediction, and so on. This research work Additionally, when compared against a
aims in proposing a better genetic standard decision tree, their method builds
algorithm based feature selection strategy extremely smaller trees in only a fifth of
based five layered artificial neural network the time, while having a superior
(IGA-FLANN) classifier. The aim of IGA- performance consistent by cost savings,
FLANN is to better the prediction leading to a method that not only has more
accuracy and also to decrease the time business-oriented results, but also a
taken for classification.

ISSN: 2395-1303 Page 200

International Journal of Engineering and Techniques - Volume 3 Issue 5, Sep - Oct 2017

method that creates simpler models that the coaching conversation with the highest
are easier to analyze. value variable. A weakness of their
approach is that it does not take into
Online decision trees from data streams consideration the communication between
are usually unable to handle concept drift. behaviour variables and, due to the
Blanco et al., 2016 proposed the uncertainty congenitally present in
Incremental Algorithm Driven by Error modelling learning styles, small
Margins (IADEM-3) that mainly carry out differences in behaviour can lead to
two actions in response to a concept drift. incorrect predictions. Subsequently, the
At first, IADEM-3 resets the variables learner is presented with coaching material
affected by the change and continue not suited to their learning style. Because
unbroken the structure of the tree, which of the above mentioned challenges a new
allows for changes in which ensuing target method that uses fuzzy decision trees to
functions are very similar. After that, build a series of fuzzy predictive models
IADEM-3 creates alternative models that connecting these variables for all
replace parts of the main tree when they dimensions of the Felder Silverman
decidedly improve the accuracy of the Learning Styles model. Results using live
model, thereby rebuilding the main tree if data by the authors showed that the fuzzy
needed. An online change detector and a models have elevated the predictive
non-parametric statistical test based on certainty across four learning style
Hoeffding’s bounds are used to agreement dimensions and promote the discovery of
that significance. A new pruning method is some interesting relationships amongst
also incorporated in IADEM-3, making behaviour variables.
sure that all split tests previously installed
in decision nodes are useful. Their learning 2.2. Recent Works on Support Vector
model is also viewed as an in concert of Machine (SVM)
classifiers, and predictions of the main and
alternative models are combined to Motivated by the KNN trick presented in
classify unlabeled examples. IADEM-3 is the weighted twin support vector machines
empirically compared with various well- with local information (WLTSVM), Pan et
known decision tree induction algorithms al., 2015 expected a novel K-nearest
for concept batch detection. The authors neighbour based structural twin support
depict that their new algorithm often vector machine (KNN-STSVM). By
reaches higher levels of accuracy with applying the intra-class KNN method,
smaller decision tree models, continue the different weights are disposed to the
processing time bounded, irrespective of samples in one class to build up the
the number of instances processed. structural information. For the other class,
the excessive constraints are deleted by the
Predicting learning styles in conversational inter-class KNN method to speed up the
intelligent coaching systems using fuzzy coaching process. For large scale
decision trees has been proposed by problems, a fast clip algorithm is further
Crokett et al., 2017. Prediction of learning introduced for increase of rate.
style is carried out by confining Comprehensive experimental results on
independent behaviour variables during twenty-two datasets determine the

ISSN: 2395-1303 Page 201

International Journal of Engineering and Techniques - Volume 3 Issue 5, Sep - Oct 2017

efficiency of their proposed KNN- use of Lagrange multipliers and duality

STSVM. theory. A fast iterative solution algorithm
based on Cholesky decomposition with
It is noteworthy that existing structural permutation of the support vectors is
classifiers do not balance structural proposed as a solution method. The
information’s relationships both intra-class properties of their SVM formulation were
and inter-class. Connecting the structural analyzed and compared with standard
information with nonparallel support SVMs using a simple example that can be
vector machine (NPSVM), D. Chen et al. illustrated graphically. The correctness and
2016, construct a new structural behaviour of their proposed work has been
nonparallel support vector machine (called determined using a set of public
SNPSVM). Each model of SNPSVM benchmarking problems for both linear
examine not only the density in both and nonlinear SVMs.
classes by the structural information but
also the reparability between classes, thus Utkin and Zhuk., 2017 proposed a well-
it can fully accomplishment prior known one-class classification support
knowledge to directly improve the vector machine (OCC SVM) dealing with
algorithms generalization capacity. interval-valued or set-valued training data.
Moreover, the authors applied the enhance Their key idea is to represent every
alternating direction method of multipliers interval of training data by a finite set of
(ADMM) to SNPSVM. Both their model precise data with estimated weights. Their
itself and the solving algorithm can representation is based on replacement of
assurance that it possibly would deal with the interval-valued expected risk produced
large-scale classification problems with a by interval-valued data with the interval-
huge number of occurrence as well as valued expected risk produced by
features. Experimental results show that imprecise weights or sets of weights. It can
SNPSVM is superior to the other current also be considered that, the interval
algorithms based on structural information uncertainty is replaced with the estimated
of data in both computation time and weight or probabilistic uncertainty. The
classification accuracy. authors showed how constraints for the
estimated weights are incorporated into
Peng et al., 2016 formulated a linear kernel dual quadratic programming problems
support vector machine (SVM) as a which can be viewed as development of
regularized least-squares (RLS) problem. the well-known OCC SVM models. With
By defining a set of signal variables of the the help of numerical examples with
errors, the solution to the RLS problem is synthetic and real interval-valued training
represented as an equation that relates the data the authors decorate their proposed
error vector to the indicator variables. approach and investigate its properties.
Through dissolution the training set, the
SVM weights and bias are expressed
analytically using the support vectors. The
authors also determine how their approach 2.3. Recent Works on Adaboost
naturally extends to sums with nonlinear Universum data usually does not belong to
kernels whilst avoiding the need to make any class of the training data, has been

ISSN: 2395-1303 Page 202

International Journal of Engineering and Techniques - Volume 3 Issue 5, Sep - Oct 2017

applied for training better classifiers. Xu et verified the performance of Rob_MulAda

al., 2014 addressed a novel boosting and provide a suggestion in choosing the
algorithm called UAdaBoost which most convenient noise-alleviating
possibly would boost the classification approach according to the concrete noise
performance of AdaBoost with Universum level in practical applications.
data. UAdaBoost chooses a function by
minimizing the loss for labelled data and Baig et al., 2017 presented a boosting-
Universum data. The cost function is based method of learning a feed-forward
minimized by a greedy, stagewise, artificial neural network (ANN) with a
functional gradient procedure. Each single layer of hidden neurons and a single
training stage of UdaBoost is fast and output neuron. At first, an algorithm called
efficient. The standard AdaBoost weights Boostron is represent which learns a
labeled samples during training emphasis single-layer perceptron using AdaBoost
while UAdaBoost gives an explicit and decision stumps. It is then continue to
learn weights of a neural network with a
weighting design for Universum samples
single hidden layer of linear neurons. At
as well. Also the authors described the
practical conditions for the capability of last, a novel method is introduced by the
Universum learning. These conditions are authors to incorporate non-linear
based on the analysis of the distribution of activation functions in artificial neural
at once predictions over training samples. network learning. Their proposed method
By their experimental results the authors uses continue representation to
claimed that their method can obtain approximate non-linearity of activation
superior performances over the standard functions, learns the coefficients of
AdaBoost by selecting proper Universum nonlinear terms by AdaBoost which adapts
data. the network parameters by a layer-wise
iterative traversal of neurons and an
Sun et al., 2016 quoted a representative appropriate reduction of the problem.
access named noise-detection based Comparison of various neural network
AdaBoost (ND_AdaBoost) in order to models learned the proposed methods and
boost the robustness of AdaBoost in the those ground using the least mean squared
two-class classification scenario. In order learning (LMS) and the resilient back-
to courage the dilemma a robust multi- propagation (RPROP) is provided by the
class AdaBoost algorithm (Rob_MulAda) authors.
is proposed by the authors whose key
ingredients consist in a noise-detection Miller and Soh 2015 proposed a novel
based multi-class loss function and a new cluster-based boosting (CBB) approach to
weight updating scheme. The authors address limitations in boosting on
claims that their experimental study supervised learning (SL) algorithms. Their
announce that their newly-proposed CBB approach partitions the training data
weight updating scheme is indeed more into clusters consist of highly similar
robust to mislabeled noises than that of member data and integrates these clusters
ND_AdaBoost in both two-class and directly into the boosting process. Their
multi-class scenarios. As well, through the CBB approach experiments to address two
comparison experiments, the authors also specific limitations for current boosting

ISSN: 2395-1303 Page 203

International Journal of Engineering and Techniques - Volume 3 Issue 5, Sep - Oct 2017

both resulting from boosting converge on To select the most feasible and compact
incorrect training data. The first one is feature subset, the improved GA with
filtering for subsequent functions when the mutation pool is being proposed in this
training data contains dangerous areas research work. The search for feature
and/or label noise; and the second one is δ
subset is done based on and conditional
over fitting in subsequent functions that mutual information measure. Every dataset
are forced to learn on all the incorrect is first fed into the proposed theoretical
instances. The authors demonstrated the search based feature selection algorithm to
capability of CBB through extensive obtain a subset of features, which are
empirical results on 20 UCI benchmark evaluated and finally after convergence,
datasets and proclaimed that CBB achieves the non-dominated feature subset is
superior predictive accuracy that use selected.
selective boosting without clusters.
As GA is a population based stochastic
III. PROPOSED WORK search algorithm, the initial population for
proposed improved GA is created at
The proposed work has contributions three
random. The dimension of population set
fold. At first an improved genetic p ×n p
algorithm based feature selection action is is where the defined population
size is and is the number of features in
portrayed. Once when the features are
selected then in the next stage, verdict the dataset. Binary string representation of
association is done. At the third stage, five chromosome is chosen in this work. The
layered artificial neural network is length of the string is same with total
proposed. number of features present in the dataset.
Each gene consists of only two values, ‘1’
3.1. Improved Genetic Algorithm based and ‘0’, which implies that the index
Feature Selection Strategy feature is present and absent respectively
P={P,P,...,P} in current subset. So, a chromosome 1
Let, 1 2 n
be the feature set
length K is represented as follows:
where n is the number of features and
U ={O 1,O2, m }∈
...,O R m× n
In the proposed improved GA, every
be a given
≤n . The population member has equal probability
dataset with m objects where m
of being a parent. Single point crossover
feature selection method can be viewed as approach is used to get offspring from
a mapping
g ,U
,D) →{P
, where,
parent chromosomes. The main reason
is the feature selection method, D is the behind application of alteration is to
decision attribute representing class labels diversity in the population. The method
'⊂P , where, P uses a alteration pool of different mutation
and P , number of
strategies. At any moment for any current
selected features, k<<n . The goal of offspring, alteration strategy to be applied
proposed method is to find P ' which will
is chosen dynamically from the pool. The
be highly relevant to D and less related to reason behind selecting such alteration
each other. strategy is to maximize quality of
population member by utilizing
effectiveness of those individual mutual

ISSN: 2395-1303 Page 204

International Journal of Engineering and Techniques - Volume 3 Issue 5, Sep - Oct 2017

strategies. To acquire the robust and noise P

9: Select another parent i
from the
resilient feature subset, the main dataset is
first sampled using sample with
replacement strategy (SWR) into E 10: Apply single point crossover
number of dataset of equal dimension with P P
between i
and i
to produce offspring
main dataset. The SWR strategy is used to
make E number of dataset from the main 11: Dynamically select mutation
dataset with intentional perturbation into it strategy from mutation pool
and reason behind choosing SWR strategy
is that in this strategy two sample values 12: Apply mutation to offspring
are independent. Goal of propose method
13: Compute two fitness values for
is to apply multiple improved GAs on
multiple sub datasets to make robust and
generalized feature subset. Among the 14: if both fitness values of offspring is
final reduct set, pair wise dominance check optimal than the global best then
is opted and at last the final selected
feature subset is that one which has its Offspring replaces one of its
domination count 0 implies it is a non- parents and update the global best
dominated solution. If multiple feature Else if Offspring dominates any
subsets are present with dominance count parent then
zero then objective wise weak dominance
concept is opted to get most feasible one. Offspring replaces that parent
The improved GA algorithm is given
End if
15: End for
Algorithm 1 Improved genetic algorithm
based feature selection (FSGA). 16: Until predefined number of
generations are exhausted
1: Input: Sampled Dataset
17: Return Non-dominated set of
2: Begin
population members as solution
3: Generate Population
) of size N
18: End
4: Evaluate both fitness functions for all 19: Output: Non-dominated reduct set
population members
3.2. Verdict Consolidation
5: Calculate global best of each fitness
function In proposed work a nondominated solution
based verdict consolidation is considered
6: Repeat for final feature subset selection. The
..N7: ensemble feature selector runs Enumber
for i=1
of individual feature selectors in parallel
ct R
8: First_parent =
i which provides set for
,2,...,E. As a result, final r

ISSN: 2395-1303 Page 205

International Journal of Engineering and Techniques - Volume 3 Issue 5, Sep - Oct 2017

is obtained by combining all such F
8: Return F
, where,
9: End
10: Output: Set F of optimal feature
Now each pair of r
ct x and y
set Ris compared based on their
objective functions and if all objective 3.3. Five Layered - ANN (FL – ANN)
function values for x are optimum compare Classifier
y y
to that of , we say that x dominates
x >y
Once after the feature selection is carried
and denote it as . For each reduct in out, FL – ANN is used to classify. FL –
R , a dominating factor is set initially as ANN is a five layered RBF based classifier
zero and it increases by one when it neural network that makes use of gradient
dominates another in R. Finally, descent approach and regression based
is selected in set F if it classification. It optimizes flattening
dominates all the reducts, i.e., if it’s parameter of RBF kernel through grade
R−1 descent approach.
dominating factor is . If multiple
is there in F then one is Applied input vector x is transmitted to
chosen at random as the final reduct. classification layer through input layer.
Pattern layer includes one neuron for each
Algorithm 2 Find optimal features by training datum with RBF kernel. Squared
verdict consolidation. Euclidean distance between input vector x
and training data vector t is calculated as
1: in (1) where p denotes total number of
Input : All generated reducts training data at pattern layer.
it(j)= x
s −tj ,1≤j≤p
F=φ 2: Begin F=φ ... (1)

Calculated squared Euclidean distances are

dx =0
3: for each
ct x in R set used in RBF kernel function as in (2)
dx =0 where r (j) denotes output of jth training
dominating factor
data and σ represents flattening
R4: for each
ct x in R parameter. Outputs of RBF kernel function
are the output values of pattern layer
R5: for each
cty(≠x) in R neurons. Moreover, this layer includes N
target values of each training datum
(x>y) then increase dx determined by corresponding class.
6: if by
one 
d t(j)

σ2 
r(j)=e 2
... (2)
} 7: if
x) then
} When a training datum belongs to ith class
then its ith value will be 0.9 and others will
8: End-for be 0.1, as given in (3).

ISSN: 2395-1303 Page 206

International Journal of Engineering and Techniques - Volume 3 Issue 5, Sep - Oct 2017

s1≤i≤N calculates denominator value the same as
e 1≤j≤p FL-ANN represented by D.
... (3) p
D r(j)
N+1 neurons are placed at summation j=1
... (6)
layer where N is the total number of
classes and additional one term to N is for Each class is represented with a neuron at
one neuron to obtain denominator. FL- normalization layer. These neurons divide
ANN uses diverge effect term at corresponding nominator value by
summation layer to increase the distances denominator value calculated at
among classes. Diverge effect term value summation layer, according to (7) where ci
is calculated as in (4) where d (j, i) denotes denotes normalized output of ith class.
diverge effect term of jth training data and
ith class. ymax is initialized to 0.9 which i=
c i
D ... (7)
denotes the maximum value of y(j, i). ymax
value is updated with the maximum value
Class of input vector is determined at
of output layer after each iteration of
output layer through the winner decision
optimization. Diverge effect term is
mechanism as given in (8) where c is the
calculated by N neurons of summation
output vector of normalization layer, cid
layer. This calculation includes aggressive
and id denote winner neuron value and
form of y (j, i) − ymax to increase the effect
indices of the class, respectively.
of y (j, i).
[c ,i
d ]=m
... (8)
... (4)
Gradient descent based collective learning
Diverge effect term is used in calculating
is utilized in FL-ANN for obtaining
nominator values at summation layer as in
optimized flattening parameter value. Each
(5). Moreover, denominator value is also
training datum at pattern layer is
calculated at this layer as in (6).
sequentially applied to neural network and
p three steps are executed until maximum
ui = ∑d ( j , i ) * r ( j ),1 ≤ i ≤ N iteration limit exceeds. Firstly, squared
... (5) error e is calculated for each input, as in
(9) where y(z, id) represents the value of
When N neurons, represented with ui,
zth training input data for idth class and cid
calculate nominator values by summing
is value of winner class.
dot product of diverge effect terms and
classification layer outputs, other neuron e=(y(
d )−c )2
... (9)

ISSN: 2395-1303 Page 207

International Journal of Engineering and Techniques - Volume 3 Issue 5, Sep - Oct 2017


Table 1. Dataset Name, No. of Instances and No. of Features

Dataset Name No. of Instances No. of Features

Blood 748 5

Bupa 345 7

Car 1728 6

Contraceptive 1473 9

Credit 30000 24

Diagnostic 569 32

Ecoli 336 8

Ionosphere 351 34

Mammography 961 6

monks – 1 432 7

monks – 2 432 7

monks – 3 432 7

Parkinsons 197 23

Pima 768 8

Prognostic 198 34

Sonar 208 60

Spect 267 22

Tic-Tac-Toe Endgame 958 9

Vert 310 6

Yeast 1484 8

ISSN: 2395-1303 Page 208

International Journal of Engineering and Techniques - Volume 3 Issue 5, Sep - Oct 2017

Table 2. Predictive Accuracy of the Algorithms


blood 0.8 0.78 0.78 0.78 0.83

bupa 0.7 0.72 0.67 0.69 0.76

car 1 0.99 0.96 0.97 1

cont. 0.72 0.69 0.7 0.68 0.77

credit 0.87 0.87 0.86 0.86 0.89

diag. 0.97 0.98 0.95 0.98 0.99

ecoli 0.99 0.99 0.99 0.98 0.99

iono. 0.91 0.91 0.89 0.94 0.96

mamm. 0.82 0.82 0.82 0.81 0.85

mks-1 1 1 0.99 0.96 1

mks-2 1 1 0.68 0.71 1

mks-3 1 1 1 1 1

park 0.92 0.94 0.9 0.9 0.97

pima 0.77 0.77 0.75 0.77 0.81

prog. 0.8 0.77 0.79 0.8 0.84

sonar 0.81 0.82 0.79 0.83 0.89

spect 0.81 0.82 0.84 0.85 0.88

Tic 0.97 1 0.94 0.97 1

vert 0.82 0.82 0.83 0.83 0.87

yeast 0.65 0.66 0.66 0.64 0.68

ISSN: 2395-1303 Page 209

International Journal of Engineering and Techniques - Volume 3 Issue 5, Sep - Oct 2017

Fig.2. Predictive Accuracy Comparison for the datasets 1 to 10

Fig.3. Predictive Accuracy Comparison for the datasets 11 to 20

ISSN: 2395-1303 Page 210

International Journal of Engineering and Techniques - Volume 3 Issue 5, Sep - Oct 2017

Table 3. Time Taken by the Algorithms for Classification (in milliseconds)


blood 2396 2093 2377 2839 1032

bupa 2436 2699 2056 2496 1948

car 2251 2117 2157 2385 1682

cont. 2157 2369 2978 2635 1732

credit 19377 19507 19563 19674 15782

diag. 7361 7962 7233 7676 3982

ecoli 2268 2894 2367 2665 1846

iono. 7588 8000 7558 7569 3901

mamm. 2875 2586 2210 2612 1738

mks-1 2039 2612 2185 2076 1936

mks-2 2329 2956 2940 2120 1726

mks-3 2024 2003 2062 2497 1639

park 4047 4646 4526 4879 2393

pima 2052 2533 2813 2992 1392

prog. 2686 2283 2208 2644 1888

sonar 8686 8681 8297 8788 5189

spect 5348 5821 5655 5455 2291

Tic 2250 2290 2574 2745 1749

vert 2869 2497 2054 2645 1638

yeast 2088 2604 2971 2222 1843

ISSN: 2395-1303 Page 211

International Journal of Engineering and Techniques - Volume 3 Issue 5, Sep - Oct 2017

Fig.4. Time Taken for Classification by the Algorithms Comparison for the datasets 1 to 10

Fig.5. Time Taken for Classification by the Algorithms Comparison for the datasets 11 to 20

20 datasets are taken from the UCI – 1, monks – 2, monks – 3, parkinsons,

machine learning repository namely blood, pima, prognostic, sonar, spect, Tic-Tac-
bupa, car, contraceptive, credit, diagnostic, Toe Endgame, vert and yeast. The dataset
ecoli, Ionosphere, mammography, monks details such as name of the dataset,

ISSN: 2395-1303 Page 212

International Journal of Engineering and Techniques - Volume 3 Issue 5, Sep - Oct 2017

number of instances and number of [2] A. C. Bahnsen, D. Aouada, B. Ottersten,

features are portrayed in the Table 1. The “Example-dependent cost-sensitive
decision trees,” Expert Systems with
performance are done using MATLAB
Applications, vol. 42, pp. 6609-6619,
tool. The system configuration is Core I3 2015.
processor with 8 GB RAM and 1 TB hard [3] F. Blanco, J. C. Ávila, G. R. Jiménez, A.
disk that runs on Microsoft Windows 8 Carvalho, A. O. Díaz, R. M. Bueno,
operating system. Implementation metrics “Online adaptive decision trees based on
concentration inequalities,” Knowledge-
such as predictive accuracy and time taken
Based Systems, vol. 104, pp. 179-194,
for classification are obtained. For better 2016.
visual depiction at first 10 datasets are [4] Crockett, A. Latham, N. Whitton, “On
involved in the simulation followed up predicting learning styles in
with next 10 datasets. It is evident from the conversational intelligent tutoring systems
using fuzzy decision trees,” International
results that the proposed IGA-FLANN
Journal of Human-Computer Studies, vol.
outperforms all the other algorithms in 97, pp. 98-115, 2017.
terms of prediction accuracy. Next, we [5] X. Pan, Y. Luo, Y. Xu, “K-nearest
compared the implementation of the neighbor based structural twin support
proposed IGA-FLANN in terms of time vector machine,” Knowledge-Based
Systems, vol. 88, pp. 34-44, 2015.
taken for classification. From that results
[6] D. Chen, Y. Tian, X. Liu, “Structural
too, it is obvious that the proposed IGA- nonparallel support vector machine for
FLANN consumes less time than that of pattern recognition,” Pattern Recognition,
all the algorithms. vol. 60, pp. 296-305, 2016.
[7] X. Peng, K. Rafferty, S. Ferguson,
V. CONCLUSIONS “Building support vector machines in the
context of regularized least squares,”
This research manuscript aims in design Neurocomputing, volume. 211, pp. 129-
and development of an improved genetic 142, 2016.
[8] V. Utkin, Y. A. Zhuk, “An one-class
algorithm based feature selection approach
classification support vector machine
based five layered artificial neural network model by interval-valued training data,”
classifier. Around 20 datasets are obtained Knowledge-Based Systems, vol. 120, pp.
from the UCI repository. Implementations 43-56, 2017.
are carried out using MATLAB tool. [9] J. Xu, Q. Wu, J. Zhang, Z. Tang,
“Exploiting Universum data in AdaBoost
Implementation metrics such as prediction
using gradient descent,” Image and Vision
accuracy and time taken for prediction are Computing, vol. 32, pp. 550-557, 2014.
taken into account to conduct the [10] B. Sun, S. Chen, J. Wang, H. Chen, “A
Implementation evaluation of the proposed robust multi-class AdaBoost algorithm for
classifier. Simulation results presents that mislabeled noisy data,” Knowledge-Based
Systems, vol. 102, pp. 87-102, 2016.
the proposed IGA-FLANN classifier
[11] M. Baig, M .M. Awais, E. M. El-Alfy,
outperforms the existing classifiers. “AdaBoost-based artificial neural network
learning,” Neurocomputing, vol. 16, pp.
REFERENCES 22 – 41, 2017.
[12] L. D. Miller and L. K. Soh, "Cluster-
[1] Y. Lertworaprachaya, Y. Yang, R. John,
Based Boosting," IEEE Transactions on
“Interval-valued fuzzy decision trees with
Knowledge and Data Engineering, vol. 27,
optimal neighbourhood perimeter,”
pp. 1491-1504, 2015.
Applied Soft Computing, vol. 24, pp. 851-
866, 2014.

ISSN: 2395-1303 Page 213