Professional Documents
Culture Documents
Neurocomputing
journal homepage: www.elsevier.com/locate/neucom
art ic l e i nf o a b s t r a c t
Article history: Intending to achieve an algorithm characterized by the quick convergence of hard c-means (HCM) and
Received 14 February 2013 finer partitions of fuzzy c-means (FCM), suppressed fuzzy c-means (s-FCM) clustering was designed to
Received in revised form augment the gap between high and low values of the fuzzy membership functions. Suppression is
13 September 2013
produced via modifying the FCM iteration by creating a competition among clusters: for each input
Accepted 4 February 2014
vector, lower degrees of membership are proportionally reduced, being multiplied by a previously set
Communicated by M. Sato-Ilic
Available online 13 April 2014 constant suppression rate, while the largest fuzzy membership grows to maintain the probabilistic
constraint. Even though so far it was not treated as an optimal algorithm, it was employed in a series of
Keywords: applications, and reported to be accurate and efficient in various clustering problems. In this paper we
Clustering
introduce some generalized formulations of the suppression rule, leading to an infinite number of new
Fuzzy c-means clustering
clustering algorithms. Further on, we identify the close relation between s-FCM clustering models
Suppressed fuzzy c-means clustering
Fuzzy c-means with improved partition and the so-called FCM algorithm with generalized improved partition (GIFP-FCM). Finally we reveal
the constraints under which the generalized s-FCM clustering models minimize the objective function of
GIFP-FCM, allowing us to call our suppressed clustering models optimal. Based on a large amount of
numerical tests performed in multidimensional environment, several generalized forms of suppression
proved to give more accurate partitions than earlier solutions, needing significantly less iterations than
the conventional FCM.
& 2014 Elsevier B.V. All rights reserved.
1. Introduction FCM reportedly creates finer partitions than HCM, it has a reduced
but still observable sensitivity to initial cluster prototypes, but it
C-means clustering algorithms belong to unsupervised classifi- converges much slower. In spite of this drawback, FCM is one of the
cation methods which group a set of input vectors into a most popular clustering algorithms [23] not only in engineering
previously defined number (c) of classes. Initially there was the studies, but also in a wide variety of sciences from climatology [22] to
hard c-means (HCM) algorithm [31], which employed the bivalent economical forecasting [24].
(crisp) logic to describe partitions. HCM usually converges quickly, There exist several reported attempts to reduce the execution time
but is considerably sensitive to initialization [1], and frequently of FCM, without causing serious damage to the partition quality. Early
gets stuck in local minima leading to mediocre partitions. solutions—developed for computers with reduced computing power—
The introduction of fuzzy logic [41] into classification theory led turned to data approximation. The first accelerated FCM algorithms
to the definition of the fuzzy partition [34], in which every input [5,25] used integer computation only. Cheng et al. [8] proposed data
vector can belong to several classes with various degrees of member- reduction based on random sampling, leading to a fast approximative
ship, and the formulation of fuzzy clustering problems [10]. The first FCM clustering. Higher speed has been also reached via data reduc-
c-means clustering algorithm that employs fuzzy partitions is the so- tion. Eschrich et al. [11] aggregated similar input vectors into a
called fuzzy c-means (FCM) proposed by Bezdek [3], which uses a weighted example. Their solution was able to speed up FCM by an
probabilistic constraint to define the fuzzy membership functions. order of magnitude. Data aggregation was employed in image seg-
mentation, to cluster gray intensity levels instead of individual pixels,
n
speeding up the execution by two orders of magnitude [36].
Corresponding author. Tel./fax: þ 40 265 262275.
E-mail addresses: lalo@ms.sapientia.ro (L. Szilágyi),
Alternately, Lázaro et al. [30] proposed a parallel hardware
szsandor72@yahoo.com (S.M. Szilágyi). implementation to the FCM algorithm and successfully applied it
1
Tel.: þ40 265 208170; fax: þ 40 265 206211. in signal processing. Kolen and Hutcheson [27] reorganized FCM to
http://dx.doi.org/10.1016/j.neucom.2014.02.027
0925-2312/& 2014 Elsevier B.V. All rights reserved.
L. Szilágyi, S.M. Szilágyi / Neurocomputing 139 (2014) 298–309 299
reduce the necessary memory storage by eliminating the partition the degree to which vector xk belongs to cluster i, m 4 1 is the
matrix. Their solution was found significantly quicker than the fuzzyfication parameter, and dik represents the distance (any inner
classical formulation of FCM, especially in case of a large amount product norm defined by a symmetrical positive definite matrix A)
of input data. Further remarkable solutions for clustering unload- between vector xk and cluster prototype vi . FCM uses a probabil-
able amount of input data were given by Hathaway and Bezdek istic partition, meaning that the fuzzy memberships of any input
[14], and Havens et al. [15]. vector xk with respect to classes satisfy the probability constraint
The suppressed fuzzy c-means (s-FCM) algorithm, introduced by c
Fan et al. [12], proposed to make a step from FCM towards HCM, by ∑ uik ¼ 1: ð2Þ
i¼1
manipulating with the fuzzy membership functions computed in each
iteration of the FCM's alternating optimization (AO) scheme. The The minimization of the objective function J FCM is achieved by
authors defined a previously set constant suppression rate α A ½0; 1, alternately applying the optimization of J FCM over fuik g with vi
which determined the behavior of the algorithm. In each iteration, fixed, i ¼ 1…c, and the optimization of J FCM over fvi g with uik fixed,
after having determined the new fuzzy membership functions for i ¼ 1…c, k ¼ 1…n [3]. During each iteration, the optimal values are
input vector with index k, denoted by u1k ; u2k ; …; uck , the algorithm deduced from the zero gradient conditions and Lagrange multi-
looks for the largest (winner) membership value uwk, with w ¼ pliers, and obtained as follows:
arg maxi fuik ; i ¼ 1; 2; …; cg, suppresses all non-winner memberships 2=ðm 1Þ
dik
by multiplication with α and raises the winner membership value uwk u⋆
ik ¼ 8 i ¼ 1…c; 8 k ¼ 1…n; ð3Þ
2=ðm 1Þ
in such a way that the probabilistic constraint is not harmed. The ∑cj ¼ 1 djk
authors remarked that setting the suppression rate α ¼ 0 makes
s-FCM behave like HCM, while α ¼ 1 reduces the algorithm to the ∑nk ¼ 1 um
ik xk
v⋆
i ¼ 8 i ¼ 1…c: ð4Þ
conventional FCM. The authors failed to explain in detail what is ∑nk ¼ 1 umik
happening in case of other values of α, they only said that the
According to the AO scheme of the FCM algorithm, Eqs. (3) and
competition created among classes leads to beneficial effects in
(4) are alternately applied, until cluster prototypes stabilize. This
both execution time and partition quality. Later, Hung et al. [19,20]
stopping criterion compares the sum of norms of the variations of
proposed some suppression rate selection schemes, and employed
the prototype vectors vi within the latest iteration with a pre-
s-FCM to segment magnetic resonance images. Hung et al. [19]
defined small threshold value ε.
reported improved segmentation results for the suppressed FCM
Hard c-means is a special case of FCM, which uses m ¼1, and
compared to the conventional one. Further modifications and applica-
thus the memberships are obtained by the winner-takes-all rule.
tions of s-FCM were proposed recently in [35,21,32,39].
Each cluster prototype will be the average of the input vectors
In a previous paper [37], we have studied in detail the competition
assigned to the given cluster.
of clusters caused by suppression. We have introduced a quasi-
learning rate (QLR), similar to the learning rate known from conven-
2.2. FCM with improved partition
tional competitive clustering algorithms [26]. This QLR gives a
mathematically precise characterization of the competition. In this
Partitions provided by FCM have an undesired property, which is
paper we will show that the QLR established in [37] allows us to
visualized in Fig. 1 for the case of a one-dimensional problem and c¼3
define a series of generalization rules for the suppression, opening the
clusters. In the proximity of the boundary between two neighbor
gate towards novel efficient and accurate clustering algorithms. We
clusters, a zero fuzzy membership value for the third cluster would be
will further show that all sorts of suppressed FCM algorithms are
appreciated. Instead of that, these fuzzy membership functions with
optimal, as they minimize a special version of the objective function
respect to any third cluster have elevated values at the boundary of
introduced by Zhu et al. in their generalized theory of FCM algorithms
the other two clusters. Due to this behavior, Hö ppner and Klawonn
with improved partition [42].
[17] called the FCM partition multimodal. They also pointed out the
The rest of this paper is structured as follows. Section 2 presents
fact that the multimodality could be suppressed by reducing the fuzzy
the necessary details of s-FCM and the competition it creates, which
exponent m, but in most cases that is not desired because it also
stand at the basis of our investigations. Section 3 introduces several
influences the fuzzyness of the algorithm. To avoid or at least suppress
types of generalized suppression rules for s-FCM, giving an analy-
this multimodality, they introduced the so-called FCM with improved
tical and graphical description of the proposed methods. Section 4
partition (IFP-FCM), which is derived from an objective function that
discusses the question of optimality of all suppressed FCM cluster-
additionally contains a rewarding term:
ing models. Section 5 reports and discusses the benchmark tests
performed with the generalized s-FCM algorithm variants. Conclu- c n n c
2 2
J IFPFCM ¼ ∑ ∑ um
ik dik ∑ ak ∑ ðuik 1=2Þ ; ð5Þ
sions are given in the last section. i¼1k¼1 k¼1 i¼1
where parameters ak are positive numbers. The second term has the
2. Related works effect of pushing the fuzzy membership values uik, i ¼ 1…c, k ¼ 1…n
towards 0 or 1, while maintaining the probabilistic constraint.
2.1. The fuzzy and hard c-means algorithms Later, Zhu et al. [42] introduced a generalized version of this
algorithm, derived from the objective function:
The conventional FCM algorithm partitions a set of object data c n n c
2 m1
J GIFPFCM ¼ ∑ ∑ um
ik dik þ ∑ ak ∑ uik ð1 uik Þ; ð6Þ
into a number of c clusters based on the minimization of a i¼1k¼1 k¼1 i¼1
quadratic objective function. The objective function to be mini-
whose optimization leads to the partition update formula
mized is defined as
ðdik ak Þ 1=ðm 1Þ
2
c n c n
J FCM ¼ ∑ ∑ um 2 m 2 u⋆
ik ¼ 8 i ¼ 1…c; 8 k ¼ 1…n: ð7Þ
ik J x k v i jjA ¼ ∑ ∑ uik dik ; ð1Þ
∑cj ¼ 1 ðdjk ak Þ 1=ðm 1Þ
2
i¼1k¼1 i¼1k¼1
where xk represents the input data (k ¼ 1…n), vi represents the Eq. (7) explains to us the behavior of GIFP-FCM: for any input
prototype or centroid value or representative element of cluster vector xk , the square of its distances measured from all cluster
i (i ¼ 1…c), uik A ½0; 1 is the fuzzy membership function showing prototypes is virtually reduced by a constant positive value ak, and
300 L. Szilágyi, S.M. Szilágyi / Neurocomputing 139 (2014) 298–309
Fig. 1. Multimodal fuzzy membership functions in a single dimensional problem, at various values of the fuzzy exponent m. The phenomenon escalates as the fuzzy
exponent grows.
ak ¼ ω minfdik ; i ¼ 1…cg;
2
ð8Þ
i
with ω A ½0:9; 0:99, thus keeping the square of all distorted distances
positive. Using ω ¼ 1 would reduce GIFP-FCM to HCM.
It is important to remark that both of the above improved
clustering models kept FCM's prototype update formula given
in Eq. (4).
Lately, these improved FCM clustering models were involved in Fig. 2. The effect of suppression: cluster wk ¼2 is the winner here, as v2k is the
several applications [6,7]. closest prototype from vector xk . The virtually reduced distance provides an
increased membership degree to the winner cluster, while all non-winner member-
ships will be proportionally suppressed.
iteration count. This was applied, for example, by Hung 3. Time and context variant suppression means to combine both
et al. [19]. previous variation models in a single suppression rule.
2. Context sensitive or data sensitive suppression means to define a
time invariant rule of suppression, which provides a dedicated Considering the fact that suppression aims at achieving a quicker
suppression rate αk to each input vector xk . convergence and provides improved partitioning compared to FCM, it
Fig. 3. Characteristics of eight suppression schemes: the evolution of the context dependent suppression rate (αk) and the quasi-learning rate (η), both plotted against the
winner fuzzy membership uw. Unless otherwise specified, these curves represent the m¼ 2 case. Presented suppression schemes are (a) gsα-FCM or conventional s-FCM [12],
having constant suppression rate α; (b) gsθ-FCM uses constant learning rate θ; (c)–(e) suppression schemes with learning rate defined as a function of winner
β=ð1 βÞ
fuzzy membership uw, namely (c) gsρ-FCM uses η ¼ 1 ρuw (left diagram plots characteristics for m¼2,3,4 cases), (d) gsβ-FCM uses η ¼ 1 uw , (e) gsκ-FCM uses
η ¼ ½1þ ð2κ 1Þ sin ðπuw Þ=2; (f)–(h) suppression schemes defined by a direct formula between μw and uw, namely (f) gsτ-FCM uses μw ¼ ðuw þ τÞ=ð1þ uw τÞ, (g) gss-FCM uses
μw ¼ usw , (h) gsξ-FCM uses μw ¼ sin ξ ðπuw =2Þ.
302 L. Szilágyi, S.M. Szilágyi / Neurocomputing 139 (2014) 298–309
is not advisable to change the suppression rule in every iteration. This variation of the learning rate against the winner fuzzy member-
is why in the following, we will exploit the possibilities of the second ship for various values of the parameter showing the definition of
generalization way, namely we will define some specific suppression the given generalized suppression rule. On the left side of each
rules and will apply them until the convergence is achieved. All figure, the variation of the obtained suppression rate αk against
algorithms proposed in this section will be called generalized sup- the winner fuzzy membership uw is displayed, under various
pressed fuzzy c-means (gs-FCM) clustering, but there will be several constraints. Although they stem from the same equation, the
types of them. suppression rules give different characteristics in this graphical
Throughout this section, we will study algorithms that apply representation.
context dependent suppression rules, which employ a dedicated
suppression rate αk for each input vector xk . The value of αk never 3.2. Direct formula between μw and uw
depends on other input vectors xl , k a l. That is why, for the sake of
simplicity, we will denote by uw (instead of uwk (or the even more According to this approach, we may formulate a direct depen-
precise uwk k , as w also depends on k)) the largest fuzzy member- dence rule between the winner fuzzy membership before and
ship function of vector, xk , before suppression. after suppression. In the general case, we may write μw ¼ gðuw Þ
Fig. 3 presents the characteristics of eight types of suppressed with g : ½0; 1-½0; 1 and gðxÞ Z x 8 x A ½1=c; 1. Eq. (9) allows us to
FCM algorithms, the first of which—shown in Fig. 3(a) is the write
conventional version introduced by Fan et al. [12]. Each of Fig. 3
μw ¼ gðuw Þ ) 1 αk þ αk uw ¼ gðuw Þ
(a)–(h) contains two subplots: on the left side the suppression rate
) αk ð1 uw Þ ¼ 1 gðuw Þ
αk is plotted against the winner fuzzy membership uw, while
1 gðuw Þ
on the right side the dependency between the quasi-learning rate ) αk ¼ : ð14Þ
η and winner fuzzy membership function uw is indicated. The 1 uw
algorithms presented in Fig. 3(b)–(h) will be described in the Let us write a few such direct relations. For example, let us
following. employ μw ¼ usw with parameter s A ½0; 1, and μw ¼ ½ sin ðπ uw =2Þξ
with parameter ξ A ½0; 1. The obtained suppression rules are
3.1. Learning rate defined as a function of the winner indicated in the last two rows of Table 1. The algorithms derived
fuzzy membership from these suppression rules will be referred to as generalized
fuzzy c-means of type s and ξ. Further on, let us derive the
In this section we will suppose that the QLR varies according generalized fuzzy c-means of type τ from the following formula
to a function of the winner membership: η ¼ f ðuw Þ, where f : inspired by the relativistic speed addition:
½0; 1-½0; 1 is a continuous function. In this case, the context uw þ τ uw þ τ
dependent suppression rate becomes μw ¼ ) 1 α k þ α k uw ¼
1 þ uw τ 1 þ uw τ
uw þ τ
1 αk ð1 mÞ=2 ) αk ð1 uw Þ ¼ 1
η ¼ f ðuw Þ ) 1 1 þ ¼ f ðuw Þ 1 þ uw τ
α k uw
1 αk 1 þ uw τ uw τ
) ¼ ð1 f ðuw ÞÞ2=ð1 mÞ 1 ) αk ¼
α k uw ð1 uw Þð1 þuw τÞ
1τ
) αk ¼ ½1 uw þ uw ð1 f ðuw ÞÞ2=ð1 mÞ 1 : ð13Þ ) αk ¼ ; ð15Þ
1 þ uw τ
For the special case of constant learning rate we use η ¼ f ðuw Þ ¼ θ
with parameter τ ranging between 0 and 1. The above formula holds
with θ A ½0; 1. A learning rate that linearly decreases with the winner
for any uw o 1. When uw ¼ 1, the suppression rate is irrelevant.
fuzzy membership can be defined by f ðuw Þ ¼ 1 ρuw with 0 r ρ r 1.
Fig. 3(f)–(h) shows the characteristics of the three different suppres-
β=ð1 βÞ
Let us further define two more such functions: f ðuw Þ ¼ 1 uw sion rules defined along direct formulas between μw and uw.
with parameter β A ½0; 1Þ and f ðuw Þ ¼ 12½1 þ ð2κ 1Þ sin ðπ uw Þ with
parameter κ A ½0; 1. Each of these functions deduces a suppression 3.3. The gs-FCM algorithm
rule from Eq. (13). The obtained suppression rules are indicated in the
first four rows of Table 1. The algorithms derived from these four So far we have introduced seven generalized suppression rules,
suppression rules will be referred to as generalized fuzzy c-means of each governed by a parameter that can take an infinite number
type θ, ρ, β, and κ. of different values. A limited number of these values (one or
Fig. 3(b)–(e) exhibits the characteristics of the above-described two of them) reduce the generated algorithm to FCM or HCM,
suppression schemes. On the right side of each figure, we have the while every other parameter value defines a new clustering model.
Table 1
Proposed generalized suppression rules.
The suppression rules introduced above are summarized in 5. Results and discussion
Table 1. Further suppression rules can be defined similar to the
ways presented in Sections 3.1 and 3.2. 5.1. Evaluation criteria
The proposed algorithm can be summarized as follows:
A series of experimental tests was performed on standard test
1. Set the number of clusters c, and the value of the fuzzy exponent data sets, which contained ground truth information for evalua-
m 4 1. tion purposes. Our aim was to find those suppression rules, which
2. Initialize cluster prototypes according to some intelligent principles provide the highest number of correct decisions and best value
(or by selecting random input vectors for each prototype). of cluster validity indexes (CVI). We employed cluster validity
3. Choose suppression rule and set the value of its parameter,
according to Table 1.
4. Compute fuzzy membership with the conventional formula of
FCM, Eq. (3). Table 2
5. For each input vector xk , find the winner cluster, set w equal to Accuracy test results using the WINE data set: for each algorithm and each fuzzy
the index of the winner cluster, and compute the suppression exponent, the average number of correct decisions (out of 178) is exhibited,
rate αk according to the suppression rule, with the correspond- maximized with respect to the suppression parameter. Algorithms are ranked
according to their observed accuracy.
ing formula from the last column of Table 1.
6. For each input vector xk , compute suppressed fuzzy member- Algorithm Fuzzy exponent m
ships with the conventional suppression formula given in
Eq. (9), using the suppression rate αk. 1.4 1.7 2.0 2.4 2.8 3.2
Fig. 4. Results of clustering the normalized WINE data set using various suppression rules. Each image plots the average number of correct decisions (out of 178 input
vectors) against the value of the suppression parameter (α, θ, ρ, β, κ, τ, s, or ξ). Each image contains three graphs, showing the performance of the given gs-FCM algorithm for
m A f1:4; 2:0; 3:2g.
L. Szilágyi, S.M. Szilágyi / Neurocomputing 139 (2014) 298–309 305
5.2. Tests on the WINE data set we use as reference, gives 170 correct decisions for fuzzy exponent
m o 1:85; as m rises beyond this threshold, the accuracy of FCM drops
The proposed generalized suppression schemes underwent quickly.
detailed numerical tests involving the 13-dimensional WINE data set All seven proposed generalized suppression schemes, together
[2], which contains 178 input vectors, grouped into three classes of a with the s-FCM proposed by Fan et al. [12], were tested in various
different cardinality. The data set also contains a label for each vector, circumstances:
which is used as ground truth at the evaluation of accuracy. The WINE
data vectors were linearly normalized in each dimension into the [0,1] 1. various values of the fuzzy exponent m A ½1:2; 3:2, covering the
interval. range recommended in [18];
The first criterion to evaluate accuracy will be the number of 2. suppression parameter ranging from 0 to 1 in steps of 0.01;
correct decisions out of 178. The conventional FCM algorithm, which 3. 500 different, randomly chosen initialization sets of the cluster
prototypes, but the same ones for each algorithm and each
parameter value. Initialization vectors in all 500 cases were chosen
randomly, but the distance between any two vectors in the same
set—in the normalized 13-dimensional space—was at least 1.
Fig. 7. WINE data clustering results: cluster validity indexes of the six most accurate suppression rules, for various values of fuzzy exponent m, plotted against the
suppression parameter of each approach. Most suppression rules have wide ranges of their suppression parameter, for which the clusters are more valid than in FCM.
306 L. Szilágyi, S.M. Szilágyi / Neurocomputing 139 (2014) 298–309
Tables 3 and 4 summarize the comparison of the accuracy proved to have relevant advantage in accuracy over s-FCM and
of each suppression rule with the performance of FCM. For each FCM as well. The highest achieved average number of correct
algorithm, the suppression parameter varied from 0 to 1 in steps of decisions was given by suppression rule of type ξ at m 3 and
0.01, and the number of correct decisions averaged along the 500 ξ 4 0:75. The 172.5 correct decisions represent approximately 30%
different random initializations was compared with the number of less misclassifications than in case of FCM in its best setting.
FCM's correct decisions. Table 3 indicates in how many cases (out of On the other hand, suppression rules of type ρ and β led to
101) these suppression rules proved more accurate than FCM at the worse accuracy that FCM, showing that not every generalized
given fuzzy exponent m. Table 4 indicates in how many cases (out of suppression rule is useful (Fig. 4).
101) these suppression rules had more correct decisions than 170, the Fig. 5 exhibits some of the suppression rules and parameter
highest score reached by FCM under any circumstances. Both these values with the highest accuracy achieved on the WINE data set.
tables rank the suppression rules according to their clustering These graphs plot the average number of correct decisions of
accuracy. the chosen algorithms against the value of the fuzzy exponent m.
The s-FCM algorithm, as proposed in [12], proved to be more FCM visibly becomes weak above m 4 2, while most of the chosen
accurate than FCM, at constant suppression rate 0:6 o α o 0:8. The gs-FCM algorithms have their strongest accuracy in this area.
difference is most relevant in case of m 4 2. Fig. 6 relates on the efficiency of the presented algorithms,
The generalized suppression schemes of type τ and s per- showing the number of necessary iterations to reach a certain
formed quite similar to s-FCM. Suppressions of type κ, θ and ξ level of convergence (ε ¼ 10 10 ). One iteration of any gs-FCM is
Table 5
Accuracy of clustering the Breast cancer data: besides the outcome of FCM, we recorded the highest number of correct decisions achieved by each algorithm, along with the
intervals where they were achieved, and where they performed better than FCM.
Correct Best Better Correct Best Better than Correct Best Better than Correct Best Better
decisions than FCM decisions FCM decisions FCM decisions than
FCM
Fig. 8. Cluster validity indexes obtained from Breast cancer data clustering: for each studied suppression rule, black curves represent the average value of CVI computed
along the valid range of fuzzy exponent m, while the gray area indicates the minimum–maximum range of each CVI.
L. Szilágyi, S.M. Szilágyi / Neurocomputing 139 (2014) 298–309 307
extremely high value of the fuzzy exponent m ¼5. For this given literature [28], mostly without being identified as such. Details
problem, s-FCM (gsα-FCM) can distinguish the three clusters if the in this matter we have reported in [38].
suppression rate is set to α o 0:718, gsτ-FCM does the same at
τ 40:209, while gsξ-FCM creates fine partition in the whole range
of its suppression parameter. This way suppression helps small 5.6. Further remarks
separated groups to form an individual cluster, instead of being
merged with other larger groups. The clustering models introduced in this paper are fully
compatible with all extensions and applications of the fuzzy c-
means algorithm, as long as the alternate optimization scheme of
5.5. Suppressed c-means partitions in image segmentation FCM is broken only by the suppression. This includes, for example,
solutions like
Willing to demonstrate the partitions and fuzzy membership
functions that we can obtain, the proposed gs-FCM variants were 1. efficient implementation via aggregation of similar data [5,36,14],
also tested for the segmentation of single channel intensity images. or without storing the partition matrix [27];
An image whose intensity histogram contained three relevant peaks 2. fuzzy c-shell variants [9,29] and derivations [4,16] for the detection
was chosen as input data. Fig. 10 shows the variety of fuzzy of clusters with certain shapes.
membership functions achieved with the tested algorithms, for
pixel intensities in range of 0–255. Fig. 10(a) and (b) shows the 6. Conclusion
outcome of FCM, and s-FCM with constant α ¼ 0:5. The largest
membership function in s-FCM always has a granted lead of 1 α In this paper we introduced several generalization schemes for
above any non-winner membership. Non-winner memberships are the suppressed fuzzy c-means algorithm. Each scheme has a
reduced proportionally and no other deformations of the curves are suppression parameter defined in the range between 0 and 1, and
present. However, with the introduction of the gs-FCM algorithms, each of these values leads to a different clustering algorithm. We
a wide variety of membership functions become available. Two demonstrated the advantages of certain suppression models in a
of them are demonstrated in Fig. 10(c) and (d). The choice of the multidimensional environment, both in terms of accuracy and
suppression rule and the value of the suppression parameter have efficiency of the partitioning. We also demonstrated how the
great influence on the shape of the curves. The multimodal generalized suppressed FCM combats the multimodality of FCM's
behavior of fuzzy membership functions is effectively reduced by fuzzy partition. We proved the correctness of the results using
each suppression rule. several cluster validity indexes. We identified existing, reportedly
The set of suppression rules introduced in this paper is hardly accurate gs-FCM clustering models employed in image processing
complete. We have only introduced a handful of suppression rules problems, which had been introduced along a completely different
using two different recipes. Anyone can define further suppression rationale. We introduced a generalized objective function using
rules, the recipe being given in Sections 3.1 and 3.2. In fact, some which we unified the theories of FCM algorithms with improved
variants of the gs-FCM algorithm are already present in the partition [17,42] and the suppressed FCM clustering models [12,37].
Fig. 10. Fuzzy membership functions given by suppressed FCM variants on a single channel intensity image: (a) FCM; (b) s-FCM at α ¼ 0:5; (c) gs-FCM of type ξ at ξ ¼ 0:65;
(d) gs-FCM of type β at β ¼ 0:75. All these curves were obtained using fuzzy exponent m¼ 3. All suppression rules subdue the multimodality of the fuzzy membership
functions.
L. Szilágyi, S.M. Szilágyi / Neurocomputing 139 (2014) 298–309 309
This novel objective function also serves as evidence for the optimality [22] R. Huth, C. Beck, A. Philipp, M. Demuzere, Z. Ustrnul, M. Cahynová, J. Kyselý,
of the suppressed FCM algorithm variants. The easily implementable O.E. Tveito, Classifications of atmospheric circulation patterns: recent
advances and applications, Ann. N.Y. Acad. Sci. 1146 (2008) 105–152.
optimal suppressed FCM variants will surely gain several applications [23] H. Izakian, A. Abraham, Fuzzy c-means and fuzzy swarm for fuzzy clustering
in the coming years. problem, Expert Syst. Appl. 38 (2011) 1835–1838.
[24] L.X. Jun, W.Y. Xin, Q.J. Guang, W.L. Hui, The application of fuzzy c-means
clustering in macro-economic forecast, in: Proceedings of the Second Inter-
national Symposium on Electronic Commerce and Security (ISECS2009,
Acknowledgments Nanchang, China), 2009, pp. 609–611.
[25] M.S. Kamel, S.Z. Selim, New algorithms for solving the fuzzy clustering
problem, Pattern Recognit. 27 (1994) 421–428.
The work of L. Szilágyi was funded by UEFISCDI under Grant [26] T. Kohonen, The self-organizing map, Proc. IEEE 78 (1990) 1474–1480.
no. PD667, Contract no. 28/05.08.2010, with the exception of [27] J.F. Kolen, T. Hutcheson, Reducing the time complexity of the fuzzy c-means
algorithm, IEEE Trans. Fuzzy Syst. 10 (2002) 263–267.
Sections 2.2 and 4, which are the products of Grant no. PD103921
[28] S. Krinidis, V. Chatzis, A robust fuzzy local information c-means clustering
funded by the Hungarian National Research Fund (OTKA). The work algorithm, IEEE Trans. Image Process. 19 (2010) 1328–1337.
of S.M. Szilágyi was funded by the project “Transnational Network [29] R. Krishnapuram, O. Nasraoui, H. Frigui, A fuzzy c spherical shells algorithm:
for Integrated Management of Postdoctoral Research in Commu- a new approach, IEEE Trans. Neural Netw. 3 (1992) 663–671.
[30] J. Lázaro, J. Arias, J.L. Martín, C. Cuadrado, A. Astarloa, Implementation of
nicating Sciences. Institutional building (postdoctoral school) a modified fuzzy c-means clustering algorithm for real-time applications,
and fellowships program (CommScie)”- POSDRU/89/1.5/S/63663, Microprocess. Microsyst. 29 (2005) 375–380.
financed under the Sectoral Operational Program Human Resources [31] S. McQueen, Some methods for classification and analysis of multivariate
observations, in: The Fifth Berkeley Symposium on Mathematical Statistics
Development 2007–2013. and Probability, 1967, pp. 281–297.
[32] A. Nyma, M. Kang, Y.K. Kwon, C.H. Kim, J.M. Kim, A hybrid technique for medical
image segmentation, J. Biomed. Biotechnol. (2012) 1–7 (Article ID 830252).
[33] N.R. Pal, J.C. Bezdek, On cluster validity for the fuzzy c-means model, IEEE
Appendix A. Supplementary material Trans. Fuzzy Syst. 3 (1995) 370–379.
[34] E.H. Ruspini, A new approach to clustering, Inf. Control 16 (1969) 22–32.
Supplementary data associated with this paper can be found in [35] M.F. Saad, A.M. Alimi, Improved modified suppressed fuzzy c-means, in:
Proceedings of the Second International Conference on Image Processing
the online version at http://dx.doi.org/10.1016/j.neucom.2014.02.027. Theory Tool and Applications (IPTA2010, Paris), 2010, pp. 313–318.
[36] L. Szilágyi, Z. Benyó, S.M. Szilágyi, H.S. Adam, MR brain image segmentation
using an enhanced fuzzy c-means algorithm, in: Proceedings of the 25th
References Annual International Conference of IEEE EMBC (Cancún), 2003, pp. 724–726.
[37] L. Szilágyi, S.M. Szilágyi, Z. Benyó, Analytical and numerical evaluation of the
suppressed fuzzy c-means algorithm: a study on the competition in c-means
[1] D. Arthur, S. Vassilvitskii, k-meansþ þ : the advantages of careful seeding, in: clustering models, Softw. Comput. 14 (2010) 495–505.
Proceedings of Symposium on Discrete Algorithm, 2007, pp. 1027–1035. [38] L. Szilágyi, Lessons to learn from a mistaken optimization, Pattern Recognit.
[2] A. Asuncion, D.J. Newman, UCI Machine Learning Repository, available at: Lett. 36 (2014) 29–35, http://dx.doi.org/10.1016/j.patrec.2013.08.027.
〈http://archive.ics.uci.edu/ml/datasets.html〉. [39] H.S. Tsai, W.L. Hung, M.S. Yang, A robust kernel-based fuzzy c-means
[3] J.C. Bezdek, Pattern Recognition with Fuzzy Objective Function Algorithms, algorithm by incorporating suppressed and magnified membership for MRI
Plenum, New York, NY, 1981. image segmentation, in: Lecture Notes in Artificial Intelligence in Medicine,
[4] J.C. Bezdek, R.J. Hathaway, N.R. Pal, Norm induced shell prototype (NISP) vol. 7530, 2012, pp. 744–754.
clustering, Neural Parallel Sci. Comput. 3 (1995) 431–450. [40] X.L. Xie, G.A. Beni, Validity measure for fuzzy clustering, IEEE Trans. Pattern
[5] R.L. Cannon, J.V. Dave, J.C. Bezdek, Efficient implementation of the fuzzy Anal. Mach. Intell. 3 (1991) 841–846.
c-means clustering algorithms, IEEE Trans. Pattern Anal. Mach. Intell. 8 (1986) [41] L.A. Zadeh, Fuzzy sets, Inf. Control 8 (1965) 338–353.
248–255. [42] L. Zhu, F.L. Chung, S. Wang, Generalized fuzzy c-means clustering algorithm
[6] A. Celikyilmaz, I.B. Türkşen, Enhanced fuzzy system models with improved with improved fuzzy partitions, IEEE Trans. Syst. Man Cybern. B. 39 (2009)
fuzzy clustering algorithm, IEEE Trans. Fuzzy Syst. 16 (2008) 779–794. 578–591.
[7] A. Celikyilmaz, I.B. Türkşen, R. Aktaş, M.M. Doganay, N.B. Ceylan, Increasing
accuracy of two-class pattern recognition with enhanced fuzzy functions,
Expert Syst. Appl. 36 (2009) 1337–1354.
[8] T.W. Cheng, D.B. Goldgof, L.O. Hall, Fast fuzzy clustering, Fuzzy Sets Syst. 93
(1998) 49–56. László Szilágyi received M.S.c in electrical engineering,
[9] R.N. Davé, K. Bhaswan, Adaptive fuzzy c-shells clustering and detection of in 1998, from Petru Maior University of Tîrgu Mureş,
ellipses, IEEE Trans. Neural Netw. 3 (1992) 643–662. Romania, and Ph.D., in 2009, from Budapest University
[10] J.C. Dunn, A fuzzy relative of the ISODATA process and its use in detecting of Technology and Economics, Hungary. Currently he is
compact well separated clusters, J. Cybern. 3 (1974) 32–57. an associate professor at Sapientia University of Trans-
[11] S. Eschrich, J. Ke, L.O. Hall, D.B. Goldgof, Fast accurate fuzzy clustering through ylvania, Tîrgu Mureş, Romania, where he teaches Image
data reduction, IEEE Trans. Fuzzy Syst. 11 (2003) 262–270. Processing and Pattern Recognition, Modeling and Simu-
[12] J.L. Fan, W.Z. Zhen, W.X. Xie, Suppressed fuzzy c-means clustering algorithm, lation, and Parallel Algorithms. His research topics
Pattern Recognit. Lett. 24 (2003) 1607–1612. include medical image processing, artificial intelligence,
[13] Y. Fukuyama, M. Sugeno, A new method of choosing the number of clusters for and unsupervised learning methods in pattern recogni-
the fuzzy c-means method (in Japanese), in: Proceedings of the Fifth Fuzzy tion. He has published two books, 10þ journal papers,
20þ book chapters and 60þ conference papers. He is
Systems Symposium, 1989, pp. 247–250.
member of IEEE and IAPR's Hungarian Section. He was
[14] R.J. Hathaway, J.C. Bezdek, Extending fuzzy and probabilistic clustering to very
awarded the János Bolyai Fellowship Prize by the Hun-
large data sets, Comput. Stat. Data Anal. 51 (2006) 215–234.
garian Academy of Sciences, in 2010.
[15] T.C. Havens, J.C. Bezdek, C. Leckie, L.O. Hall, M. Palaniswami, Fuzzy c-means
algorithms for very large data, IEEE Trans. Fuzzy Syst. 20 (2012) 1130–1146.
[16] F. Hö ppner, Fuzzy shell clustering algorithms in image processing: fuzzy
c-rectangular and 2-rectangular shells, IEEE Trans. Fuzzy Syst. 5 (1997)
599–613. Sándor M. Szilágyi received M.Sc. in electrical engi-
[17] F. Höppner, F. Klawonn, Improved fuzzy partitions for fuzzy regression models, neering, in 1996, from Petru Maior University of Tîrgu
Int. J. Approx. Reason. 32 (2003) 85–102. Mureş, Romania, and Ph.D., in 2008, from Budapest
[18] M. Huang, Z. Xia, H. Wang, Q. Zeng, Q. Wang, The range of the value for the University of Technology and Economics, Hungary.
fuzzifier of the fuzzy c-means algorithm, Pattern Recognit. Lett. 33 (2012) Currently he is an associate professor at Petru Maior
2280–2284. University of Tîrgu Mures, Romania, where he teaches
[19] W.L. Hung, M.S. Yang, D.H. Chen, Parameter selection for suppressed fuzzy Software Engineering, Low-level Programming, and
c-means with an application to MRI segmentation, Pattern Recognit. Lett. 27 Biomedical Systems. His research topics include ECG
(2006) 424–438. signal processing and dynamic heart modeling, and
[20] W.L. Hung, Y.C. Chang, A modified fuzzy c-means algorithm for differentiation various applications of pattern recognition. He has
in MRI of ophtalmology, in: Lecture Notes in Computer Science (MDAI'06, published one book, 10þ journal papers, 15þ book
Tarragona), vol. 3885, 2006, pp. 340–350. chapters and 50þ conference papers. He was awarded
[21] W.L. Hung, D.H. Chen, M.S. Yang, Suppressed fuzzy-soft learning vector János Bolyai Fellowship Prize by the Hungarian Acad-
quantization for MRI segmentation, Artif. Intell. Med. 52 (2011) 33–43. emy of Sciences, in 2011.