Professional Documents
Culture Documents
Neural Networks
journal homepage: www.elsevier.com/locate/neunet
removed. The facial feature image can be thus obtained after re- 3. Data reprocessing for facial feature extraction
moving the unnecessary features by augmenting the conven-
tional recognition system by the ASM process. That is, facial In this section, we briefly elaborate on the histogram equaliza-
features are extracted by removing the background and obstacles tion, AdaBoost algorithm, ASM and PCA as they are being used at
from input images by using the ASM. the phase of data preprocessing.
In the sequel, face extraction and recognition are carried out
by running a series of algorithms such as ASM, PCA, and DE-based 3.1. Histogram equalization
P-RBF NNs. The main objective of this study is to improve the face
recognition rate by handling the image of enhanced facial features Histogram equalization (HE) is a commonly used technique for
through the multi-dimensional data preprocessing technologies enhancing image contrast (Gonzalez & Woods, 2002).
(PCA combined with ASM) and DE-based P-RBF NNs. We realize Consider a facial image W with N pixels, and a total number of
this objective within the context of a practical recognition environ- k gray levels, e.g., 256 gray levels in the dynamic range of [0, L 1].
ment. Along with this new design approach we provide a thorough The generic idea is to map the gray levels based on the probability
comparative analysis contrasting the performance of the recog- distribution of the image input gray levels. For a given image W ,
nition scheme proposed here with the performance of other ap- the probability density function (PDF) of PDF (Wk ) is defined as
proaches existing in the literature.
This paper is organized as follows: in Section 2, we introduce the PDF (Wk ) = nk /n. (1)
proposed overall face recognition system and the design method.
For k = 0, 1, . . . , L 1, where nk represents the number of
Section 3, Histogram equalization, AdaBoost, ASM, and PCA
times that the level, Wk appears in the input image W , and n is the
forming the preprocessing part of face recognition. Optimization
total number of samples in the input image. Note that PDF (Wk ) is
techniques and a design method of a pattern classifier for face
associated with the histogram of the input image which represents
recognition are covered in Section 4. In Section 5, we analyze the
the number of pixels that have a specific input intensity Xk . A plot
performance of the proposed system by using input images data
of nk versus Wk is known as the histogram of image W (I , J ). Based
coming from a CCD camera. Finally, the conclusions are covered in
on the PDF, the cumulative density function (CDF) is defined as
Section 6.
k
2. Structure of face recognition system CDF (w) = PDF (Wj ) (2)
j =0
In this section, an overall structure of proposed face recog-
nition system and design method is described. Face recognition where Wk = w , for k = 0, 1, . . . , L1. By definition, PDF (WL1 ) =
system is constructed to data preprocessing and RBF NNs 1. Through histogram equalization we map the input image into
pattern classifier applying optimization techniques. Histogram the entire dynamic range (W0 , WL1 ).
equalization, AdaBoost, ASM, and PCA realized a phase of data pre- Fig. 2(a) and (b) show a face image along with its equivalent his-
processing. Histogram equalization compensates for illumination- togram. The output image produced after histogram equalization is
distortion of images. The face area is detected by using AdaBoost. given in Fig. 2(c) and (d). This result demonstrates the performance
By using the ASM, face information is extracted and eventual obsta- of the histogram equalization method in enhancing the contrast of
cles are removed. The features of face images are extracted with the an image through dynamic range expansion.
use of the PCA method. The RBF NN pattern classifier is constructed.
The classifier comprises three functional modules. These modules 3.2. AdaBoost-based face detection
realize the condition, conclusion and aggregation phase. The input
space of the condition phase is represented by fuzzy sets formed The purpose of face detection is to locate faces present in still
by running the FCM clustering algorithm (Tsekourasa, Sarimveisb, images. This has long been a focus of computer vision research
Kavaklia, & Bafasb, 2005). The conclusion phase involves a certain and has achieved a great deal of success (Gao, Pan, Ji, & Yang,
polynomial. The output of the network is determined by running 2012; Lopez-Molina, Baets, Bustince, Sanz, & Barrenechea, 2013;
fuzzy inference. The proposed RBF NNs come with the fuzzy infer- Rowley, Baluja, & Kanade, 1998; Sung & Poggio, 1998; Viola & Jones,
ence mechanism constructed with the use of the fuzzy rule-based 2004). Comprehensive reviews are given by Yang, Kriegman, and
network. In the construction of the classifier, Differential Evolution Ahuja (2002), and Zhao, Chellappa, and Phillips (2003). Among
(DE) is used to optimize the momentum coefficient, learning rate, face detection algorithms, the AdaBoost (Freund & Schapire, 1995)
and the fuzzification coefficient used in the FCM algorithm. Fig. 1 based method proposed by Viola and Jones (2001) has gained
portrays an overall architecture of the system. great popularity due to a high detection rate, low complexity,
S.-H. Yoo et al. / Neural Networks 69 (2015) 111125 113
The training shapes are all aligned by translation, rotation and where gij , k = 0, . . . , 2n + 1, is the gray-level intensity of a
scaling for minimizing the sum of squared distances between their corresponding pixel. The derivative profile of gij of length of 2n is
corresponding landmark points. Then, the mean shape x and the given as follows:
deviation of each training shape from the mean are calculated.
Principal component analysis (PCA) is then applied to capture dgij = [gij1 gij0 , gij2 gij1 , . . . , gij(2n+1) gij(2n) ]. (8)
most of the shape variations. Therefore, a shape model can be
approximated as follows: The normalized derivative profile is given by
x x + Pb (5) dgij
yij = (9)
where the P = (p1 , p2 , . . . , pt ) is the matrix whose columns are 2n
the first t eigenvectors with the largest eigenvalues arranged in a |dgijk |
descending order, and b = (b1 , b1 , . . . , bt )T is a weight vector for k=0
Each face differs from the average by the vector With this analysis, the calculations are greatly reduced, from the
order of the number of pixels in the images (N 2 ) to the order of
i = i . (13)
the number of images in the training set (M). In practice, the train-
This set of highly-dimensional vectors is then subject to prin- ing set of face images will be relatively small (M N 2 ), and the
cipal component analysis, which seeks for a set of M orthonormal calculations become quite manageable. The associated eigenvalues
vectors, Um , which the best describe the distribution of the data. allow us to rank the eigenvectors according to their usefulness
116 S.-H. Yoo et al. / Neural Networks 69 (2015) 111125
Fig. 8. Eigenface.
when characterizing the variation among the images. The eigen- Fig. 9. General architecture of RBF neural networks.
face images calculated from the eigenvectors of L span a basis set
that can be used to describe face images (Slavkovic & Jevtic, 2012).
The network exhibits a single hidden layer. Each node in the
We evaluated a limited version of this framework on an ensem-
hidden layer determines a level of activation of the receptive field
ble of 115 images (M = 115) images of Caucasian males digitized
(radial basis function) (x) given some input x. The jth output
in a controlled manner, and found that 40 eigenfaces (M = 40)
yj (x) is a weighted linear combination of the activation levels of
were sufficient to realize a very good description of face images.
the receptive fields:
In practice, a lower value of M can be sufficient for identification,
since accurate reconstruction of the image is not an absolute re- c
quirement. In the framework of face recognition, we are concerned yj (x) = wji i (x) (21)
with a pattern recognition task rather than image reconstruction. i =1
The eigenfaces span an M dimensional subspace of the original N 2
j = 1, . . . , s where s stands for the number of outputs (being
image space and hence, the M significant eigenvectors of the L ma-
equal to the number of classes encountered in a given classification
trix with the largest associated eigenvalues, are sufficient for reli-
problem). In the case of the Gaussian type of RBFs, we have
able representation of the faces in the face space characterized by
the eigenfaces.
x vi 2
Then a new face image ( ) is transformed into its eigenface i (x) = exp (22)
components (projected onto face space) in the following manner 2i2
n n
n
(i) Constant;
Quadratic; fi (x) = ai0 + aij xj + aijk xj xk . (35)
j =1 j=1 k=1 aT = [a10 , . . . , ac0 ], fx = [u1 , . . . , uc ]T .
These functions are activated by the corresponding entries of the (ii) Linear;
partition matrix and lead to local regression models located at the
aT = [a10 , . . . , ac0 , a11 , . . . , ac1 , . . . , acn ]
condition phase of the individual rules.
In case of the quadratic function, the dimensionality of the fx = [u1 , . . . , uc , u1 x1 , . . . , uc x1 , . . . , uc xn ]T .
problem increases quite quickly especially when dealing with (iii) Quadratic;
problems of high dimensionality. The reduced quadratic function
is also discussed with intent of reducing computational burden. aT = [a10 , . . . , ac0 , a11 , . . . , ac1 , . . . , acn , . . . , acnn ]
n n fx = [u1 , . . . , uc , u1 x1 , . . . , uc x1 , . . . , uc xn , . . . , uc xn xn ]T .
Reduced Quadratic; fi (x) = ai0 + aij xj + aijk x2k . (36) (iv) Reduced quadratic;
j =1 k=1
aT = [a10 , . . . , ac0 , a11 , . . . , ac1 , . . . , acn , . . . , acnn ]
The use of some constant in (33) (which are special cases of the
polynomial functions) reduces the P-RBF NNs to the standard RBF fx = [u1 , . . . , uc , u1 x1 , . . . , uc x1 , . . . , uc xn , . . . , uc x2n ]T .
neural networks as illustrated in Fig. 9.
For the discriminant function coming in the form of (40), a two-
class classifier implements the decision rule expressed by (39).
4.2.3. Aggregation phase of networks
Namely, x is assigned to 1 if the inner product aT fx is greater
Let us consider the P-RBF NNs structure by considering the
than zero and to 2 otherwise. The equation g (x) = 0 defines the
fuzzy partition realized in terms of FCM as shown in Fig. 10. The
decision surface that separates the two classes. Otherwise, a multi-
node denoted by is realized as a product of the corresponding
class classifier implements the decision rule expressed by (38).
fuzzy set and the polynomial function. The family of fuzzy sets Ai
Each output node generates a discriminant function corresponding
forms a partition (so that the sum of membership grades sum up to each class. If the ith output is larger than all remaining outputs,
to one at each point of the input space). The neuron realizes
the pattern x is assigned to ith class.
a sum as shown in (25). The output of P-RBF NNs can be obtained
by following a standard inference mechanism used in rule-based
systems (Oh, Pedrycz, & Park, 2004), 4.4. Optimized process of P-RBF NNs by using differential evolution
ui fji (x)
c c
The algorithm of Differential Evolution (DE) (Dervis & Selcuk,
yj = gj (x) = c
= ui fji (x) (37) 2004; Storn, 1997) introduced by Storn (1997) is a parallel
i=1 uk i=1
direct search method, which utilizes NP parameter vectors as a
k=1
population for each generation G. DE can be categorized into a class
where, ui = Ai (x). All the entries sum up to 1 as indicated by (9). of floating-point encoded, evolutionary optimization algorithms.
gj (x) describes here the discriminant function for discerning jth Currently, there are several variants of DE. The particular variant
class. used throughout this investigation is the DE/rand/1/bin scheme.
Based on the local polynomial-like representation, the global This scheme will be discussed here and more detailed descriptions
characteristics of the P-RBF NNs result through the composition of are provided. Since the DE algorithm was originally designed to
their local relationships. work with continuous variables, the optimization of continuous
problems is discussed first.
4.3. The discriminant function polynomial-based radial basis function The DE algorithm is a population-based algorithm using three
neural networks classifiers operators: crossover, mutation, and selection. Several optimization
parameters require tuning. These parameters are put together
There are many different ways to describe pattern classifiers. under the common name of control parameters. In fact, there
One of the most useful ways is the one realized in terms of a set of are only three real control parameters of the algorithm, namely
discriminant functions gi (x), i = 1, . . . , m (where m stands for the differentiation (or mutation) constant F, crossover constant CR, and
number of classes). The classifier is said to assign an input vector x size of population NP. The rest of the parameters are the dimension
to class i if of problem D that scales the difficulty of the optimization task;
gi (x) > gj (x) for all j = i. (38) maximum number of generations (iterations) GEN, which may
serve as a stopping condition; and the boundary constraints
Thus, the classifiers are viewed as networks that compute m
imposed on the variables that limit the feasible area.
discriminant functions and select the category corresponding to
the largest value produced by these functions. Regarding DE variants Price et al. present a notation to identify
In this paper, the proposed P-RBF NNs classifier is used for two- different ways to generate new vectors on DE.
class or multi-class problems. If a classification problem is multi- The most popular of them is called DE/rand/1/bin, where the
class one, then we use (38) as the discriminant function, otherwise, first term means differential evolution, the second term indicates
we consider the following decision rule defined commonly as a how the base vector is chosen (in this case, vector chosen at
single discriminant function g (x) in a two-class problem. random). The number in the third term expresses how many vector
differences will contribute in the differential mutation (one pair
Decide 1 if g (x) > 0; otherwise decide 2 . (39) in this case). Finally, the fourth term shows the type of crossover
The final output of networks, (37) is used as a discriminant utilized (bin from binomial in this variant).
function g (x) and can be rewritten in a form of the linear The DE algorithm can be outlined shortly as the following
combination sequence of steps:
Step 1: Set up the values of the parameters of the method
g (x) = aT fx (40)
such as a crossover rate (CR), scaling factor (SF), and mutation
where a is a vector of coefficients of polynomial functions used in type (one out of the 5 types available) and then randomly gen-
the conclusion phase of the rules in (33)(36) and fx is a matrix of erate NP population in search space. Each variable (or vec-
u. These can be defined for each of polynomials as follows. tor) in the n-dimensional search space will be denoted by the
S.-H. Yoo et al. / Neural Networks 69 (2015) 111125 119
Fig. 15. Average face obtained after running the PCA. 5.1. The Yale dataset
The dataset of faces obtained in this way consists of 350 images The Yale face dataset consists of 165 grayscale images of 15
with 10 images per person (including in total 35 persons). Fig. 14 individuals represented in the GIF format. There are 11 images per
provides an example of the extracted facial image dataset. person, one per different facial expression or configuration: center-
The weights of new facial database are calculated through light, w/glasses, happy, left-light, wearing/no glasses, normal, right
running the PCA algorithm. Fig. 15 presents an average face of 15 light, sad, sleepy, surprised, and wink. Because of the changes
persons obtained in this way. to extreme facial expressions, images concerning surprised facial
Fig. 16 shows reconstructed eigenfaces in face space by expressions are not included in the experiments. As a result, we
extracting eigenvectors corresponding to the largest eigenvalues arrive at a dataset comprising 150 images of 15 persons with 10
coming from the overall 350 eigenvectors. images per person.
The PCA weights associated with each candidate image are used Table 1 summarizes the values of the parameters of the
to form the discrimination functions associated with the RBF NNs. proposed classifier along with the parameters of the design
The weights of the candidate image are obtained in real-time. environment.
Fig. 17 is used as the inputs of the discriminant function. Each experiment we split data into 50%30%20% for training,
Fig. 17 shows discrimination process of recognition candidate validation, and testing subsets, namely, 80% (50%30% training
person based on real output by figure. As shown, we extracted and validation set) of the whole pattern is selected randomly for
the 5 dimensional feature by PCA. It means that we extracted training and the remaining pattern is used for testing purpose.
Table 1
Values of the parameters used in the experiments.
RBF NNs
Data split Training: Validation: Testing = 50%: 30%: 20%
Optimization algorithm DE
The number of evaluations of the objective function 1000 (20 50)
(Generations/swarms)
The number of rules 2, 3, 4, 5
Search space Polynomial type Linear, and reduced quadratic type
Fuzzification coefficient [1.1, 3.0]
Table 2
Classification rate obtained for the Yale dataset (without obstacle factors). The best results are shown in boldface.
Number of rules Case 1 (without ASM) Case 2 (with ASM)
L-RBF NNs RQ-RBF NNs L-RBF NNs RQ-RBF NNs
Table 3
Results for the IC&CI Lab. data (without obstacle factors). The best results shown in boldface.
Number of rules Case 1 (without ASM) Case 2 (with ASM)
L-RBF NNs RQ-RBF NNs L-RBF NNs RQ-RBF NNs
There was no overlap between the training, validation, and testing 5.2. IC&CI Laboratory dataset
sets (Lewrence, Giles, Tsoi, & Back, 1997). For each combination
of the parameters, the experiment was repeated five times. The IC&CI (Intelligent Control & Computational Intelligence) face
results are reported by presenting the average and standard database contains a set of face images taken by students in the
deviation of the classification rate obtained over these five University of Suwon. There are 10 different images of each of 35
repetitions of the experiment. The number of rules, polynomial distinct subjects. Each image was digitized and presented in the
type and the fuzzification coefficient are optimized by the DE. 200 200 pixel array with gray levels. Images feature frontal view
The experimental results expressed in terms of the classifica- faces with different facial expressions, and occlusions. The values
tion rate and its standard deviation are reported in Table 2. In of the parameters of the experimental setup are the same as used
Tables 25, the abbreviations L-RBF NNs and RQ-RBF NNS refer in the experiments with the Yale data.
to polynomial types of each RBF NNs such as linear and reduced The obtained results are included in Table 3.
quadratic type. Next, the performance of the classifier face recognition is
When the polynomial type is linear and the number of rules is reported in presence of various obstacle factors.
set to 4, the recognition rate is higher than 85% in Case 1 (without In this study, the classifier is trained by images used in the
ASM). When the polynomial type is linear and the number of rules previous experiment (Case 1). The classification rate is evaluated by
is 3, we obtain a similar recognition rate in Case 2 (with ASM). using test images that are affected by obstacles. The experiments
Notably, the recognition rate in Case 1 is slightly higher than the are completed under the same conditions as set up in the previous
one obtained for Case 2. experiments.
S.-H. Yoo et al. / Neural Networks 69 (2015) 111125 123
Table 4
Results for the IC&CI Lab. dataset (in case of wearing a cap). The best results shown in boldface.
Number of rules Case 1 (without ASM) Case 2 (with ASM)
L-RBF NNs Q-RBF NNs L-RBF NNs Q-RBF NNs
Table 5
Results for the IC&CI Lab. dataset (in case of using a mobile phone).
Rules number Case 1 (without ASM) Case 2 (with ASM)
L-RBF NNs RQ-RBF NNs L-RBF NNs RQ-RBF NNs
Fig. 18. A real-time acquired testing image for recognition (wearing a cap).
Fig. 19. A real-time acquired testing image for recognition (in case of using a mobile phone).
In case of wearing a cap (see Fig. 18). candidate with the fewest errors. Finally, recognition performance
In case of wearing a cap, Table 4 shows the experimental results is obtained by using test image and determined database image.
obtained for the two cases. Recognition performance is represented to recognition rate (%) and
When wearing a cap, the decrease of overall performance was the number of false recognition (false recognition/test data).
noticeable in comparison to the results produced in the previous The summary of the final classification results is presented in
experiment (without obstacles). The recognition rate of Case 2 is Fig. 20.
better than the one reported in Case 1 by using the effective facial When we used the non obstacle (general) images, the
features because the unnecessary image parts have been removed recognition rate of Case 1 (without ASM) is slightly better than
with the use of the ASM. Case 2 (with ASM). When obstacles are involved, the recognition
In case of using a mobile phone (see Fig. 19). rate of Case 2 is higher than the one reported for Case 1 by using
The corresponding results are reported in Table 5. the effective facial features because the unnecessary image parts
In this case, the recognition rate for Case 2 is better than Case 1 are removed by using ASM. Also, the recognition rate of RBFNNs is
because the unnecessary image parts have been removed with the better than the one reported for the PCA.
aid of ASM.
6. Conclusions
5.3. Comparative analysisPCA and RBF NNs
In this study, the proposed face recognition system comprises
We now compare recognition performance of PCA and RBF NNs two main functional modules. In the preprocessing part, two-
in order to evaluate the performance of the proposed model. The dimensional gray face images are obtained by using AdaBoost and
process of face recognition experimental using PCA is given as then histogram equalization is used to improve the quality of
follows. image. The personal profile consists of the extracted face contour
First, the data are divided into training data (50%), validation and shape by using the ASM. The features were extracted by the
data (30%), and testing data (20%) exactly in the same conditions PCA algorithm. In the classifier part, we proposed the optimized
as for the RBF NNs. Database consists of PCA weights of training RBF NNs for the problem of face recognition. In the recognition
images. Next, we calculate Euclidean distance between test part, the proposed P-RBF NNs exhibit some unique and useful
image weights and database image (Yale and IC&CI Lab. image) characteristics. The P-RBF NNs involve a partition module formed
weights. Then the final recognized candidate is decided as the by the FCM clustering and used here as an activation function
124 S.-H. Yoo et al. / Neural Networks 69 (2015) 111125
of the neurons located in the hidden layer. The P-RBF NNs are Dervis, K., & Selcuk, O. (2004). A simple and global optimization algorithm for
expressed as a collection of if-then fuzzy rules. The image engineering problems: Differential evolution algorithm. Turkish Journal of
Electrical Engineering, 12, 5360.
data obtained from the CCD camera is used for preprocessing Freund, Y., & Schapire, R. E. (1995). A decisiontheoretic generalization of on-line
procedures including image stabilizer, face detection and feature learning and an application to boosting. In European conference on computational
extraction. The preprocessed image data are processed with the learning theory (pp. 2327).
Gao, Y., Pan, J., Ji, G., & Yang, Z. (2012). A novel two-level nearest neighbor
aid of RBF NNs. We have reported recognition rates of the ASM classification algorithm using an adaptive distance metric. Knowledge-Based
processed images, generic images and facial images containing Systems, 26, 103110.
some obstacles. In the presence of obstacles, the recognition rate Gonzalez, R. C., & Woods, R. E. (2002). Digital image processing. New Jersey: Prentice-
of Case 2 is better than the one reported in Case 1 by using the Hall.
Han, H. G., Chen, Q. L., & Qiao, J. F. (2011). An efficient self-organizing RBF neural
effective facial features. This improvement can be attributed to the network for water quality prediction. Neural Networks, 24, 717725.
fact that the unnecessary parts of the image have been removed Huang, Y.-S., Hzu, T.-C., & Cheng, F.-H. (2010). Facial landmark detection by
combining object detection and active shape model. In Electronic commerce and
with the use of the ASM.
security, ISECS (pp. 381386).
Lewrence, S., Giles, C. L., Tsoi, A. C., & Back, A. D. (1997). Face recognition: A
Acknowledgment convolutional neural network approach. IEEE Transactions on Neural Networks,
8(1), 98113.
Lienhart, R., & Maydt, J. (2002). An extended set of Haar-like features for rapid
This work was supported by the GRRC program of Gyeonggi object detection. In IEEE conference on image processing, ICIP, New York (1)
province [GRRC Suwon 2015-B2, Center for U-city Security & (pp. 900903).
Surveillance Technology]. Lopez-Molina, C., Baets, B. D., Bustince, H., Sanz, J., & Barrenechea, E. (2013).
Multiscale edge detection based on Gaussian smoothing and edge tracking.
Knowledge-Based Systems, 50, 101111.
References Lu, X., & Jain, A. K. (2005). Multimodal facial feature extraction for automatic
3D face recognition, technical report MSU-CSE-0522. Computer Science and
Balasubramanian, M., Palanivel, S., & Ramalingam, V. (2009). Real time face and Engineering, Michigan State University.
mouth recognition using radial basis function neural networks. Expert Systems Mohammed, A. A., Minhas, R., Jonathan Wu, Q. M., & Sid-Ahmed, M. A. (2011).
with Applications, 36, 68796888. Human face recognition based on multidimensional PCA and extreme learning
Bezdek, J. C. (1981). Pattern recognition with fuzzy objective function algorithms. New machine. Pattern Recognition, 44(1011), 25882597.
York: Plenum Press. Oh, S.-K., Kim, W.-D., Pedrycz, W., & Park, H.-S. (2014). Fuzzy radial basis
Boehnen, C., & Russ, T. (2005). A fast multi-modal approach to facial feature function neural networks with information granulation and its parallel genetic
detection. In Proceedings of the seventh IEEE workshop on applications of computer optimization. Fuzzy Sets and Systems, 237, 96117.
vision, WACV/MOTION05 (pp. 135142). Oh, S.-K., Pedrycz, W., & Park, B.-J. (2004). Self-organizing neurofuzzy networks in
Chellappa, R., Wilson, C. L., & Sirohey, S. (1995). Human and machine recognition of modeling software data. Fuzzy Sets and Systems, 145, 165181.
faces: A survey. Proceedings of the IEEE, 83(5), 704740. Park, B.-J., Oh, S.-K., & Kim, H.-K. (2008). Design of polynomial neural network
Colbry, D., Stockman, G., & Jain, A. (2005). Detection of anchor points for 3D face classifier for pattern classification with two classes. Journal of Electrical
verification. In Proc. of A3DISS, San Diego CA. Engineering & Technology, 3(1), 108114.
Colombo, A., Cusano, C., & Schettini, R. (2005). 3D face detection using curvature Roh, S.-B., Oh, S.-K., & Pedrycz, W. (2010). A fuzzy of parallel polynomial neural
analysis. In 4 Intl symposium on. image and signal processing and analysis. ISPA. networks with information granules formed by fuzzy clustering. Knowledge-
Connolly, J.-F., Granger, E., & Sabourin, R. (2012). Evolution of heterogeneous Based Systems, 23(3), 202219.
ensembles through dynamic particle swarm optimization for video-based face Rowley, H. A., Baluja, S., & Kanade, T. (1998). Neural network-based face detection.
recognition. Pattern Recognition, 45(7), 24602477. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(1), 2338.
Cootes, T., Cooper, D., Taylor, C., & Graham, J. (1995). Active shape models-their Slavkovic, M., & Jevtic, D. (2012). Face recognition using eigenface approach. Serbian
training and application. Computer Vision and Image Understanding, 61(1),
Journal of Electrical Engineering, 9(6), 121130.
3839.
Cootes, T. F., Taylor, C. J., Lanitis, A., Cooper, D. H., & Graham, J. (1993). Building and Staiano, A., Tagliaferri, R., & Pedrycz, W. (2006). Improving RBF networks
using flexible models incorporating grey-level information. In Fourth internat. performance in regression tasks by means of a supervised fuzzy clustering.
conf. computer vision (pp. 242246). Neurocomputing, 69, 15701581.
S.-H. Yoo et al. / Neural Networks 69 (2015) 111125 125
Storn, R. (1997). Differential evolution, a simple and efficient heuristic strategy for Viola, P., & Jones, M. J. (2004). Robust real-time object detection. International
global optimization over continuous spaces. Journal of Global Optimization, 11, Journal of Computer Vision, 57(2), 137154.
341359. Wang, Qihui, Xie, Lijun, Zhu, Bo, Yang, Tingjun, & Zheng, Yao (2013). Facial
Sung, K.-K., & Poggio, T. (1998). Example-based learning for view-based human face feature extraction based on active shape model. Journal of Multimedia, 8(6),
detection. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(1), 747754.
Yang, M.-H., Kriegman, D., & Ahuja, N. (2002). Detecting faces in images: a survey.
3951. IEEE Transactions on Pattern Analysis and Machine Intelligence, 24(1), 3458.
Tsekourasa, G., Sarimveisb, H., Kavaklia, E., & Bafasb, G. (2005). A hierarchical fuzzy
Yong, L., Zhang, C. S., Xiaoguang, L., & Zhang, D. (2000). Face contour extraction with
clustering approach to fuzzy modeling. Fuzzy Sets and Systems, 150(2), 245266. active shape models embedded knowledge. In Internat. conf. signal processing 2,
Viola, P., & Jones, M. J. (2001). Rapid object detection using a boosted cascade of (pp. 13471350).
simple features. In IEEE conference on computer vision and pattern recognition, Zhao, W., Chellappa, R., & Phillips, P. (2003). Face recognition: a literature survey.
CVPR, Hawaii (1) (pp. 511518). ACM Computing Surveys, 35(4), 399458.