On A Type-2 Fuzzy Clustering Algorithm: Leehter Yao and Kuei-Sung Weng
On A Type-2 Fuzzy Clustering Algorithm: Leehter Yao and Kuei-Sung Weng
I.
INTRODUCTION
ISBN: 978-1-61208-221-9
45
PATTERNS 2012 : The Fourth International Conferences on Pervasive Patterns and Applications
ISBN: 978-1-61208-221-9
Figure 1. The points have the same membership value but have different
possibility values
t = u A ( i) ,
(3)
pik 1
)2
pik
membership function e
.
The confidence intervals for varying possibilistic values
pik built around the same prototype datum xik with
46
PATTERNS 2012 : The Fourth International Conferences on Pervasive Patterns and Applications
tik = ik A tik = ik e
1 p 1
ik
2 pik
i =1 k =1
JT 2GKA (T,V, A ) =
(5)
where tik (ik) denotes the Type-2(1) memberships, pik
denotes the membership degrees for one datum resembling
the possibility of its being a member of the corresponding
cluster. For example, for the Type-1 membership
value = 0.5 , the following evaluations process interprets
that Type-2 fuzzy sets evaluate their secondary membership
values with different possibility values. The prototype data
points xk, k =1,,N, have Type-1 membership value
ik = 0.5 and possibility value pik = 1.0 then the Type-2
membership values tik = 0.5 are obtained using (5). For the
same Type-1 membership value ik = 0.5 , and possibility
value pik = 0.1 we obtain the Type-2 membership values as
tik = 1.2884e 018 0 .
We know that in our design the secondary membership
function, for the same Type-1 membership value, a larger
possibility value can make the Type-1 membership value
larger than the smaller possibility value does. Using the
aforementioned concepts, we combined the probability and
possibility membership values and propose the Type-2
Gustafson-Kessel Algorithm (T2GKA).
III.
2
(tik )m DikA
+
i
( A
i
i =1
( t
k
ik
k =1
i )
(6)
1),
i =1
(8)
F =
i
N
k =1
Fi 1 i = 1, , c,
( tik )m ( xk vi )( xk vi )T
m
k =1 ( tik )
N
(10)
(11)
Fi = (1- ) Fi + det( F0 )
I,
(12)
(t ) x
=
(t )
m
vi
k =1 ik
N
(13)
k =1 ik
ISBN: 978-1-61208-221-9
tik =
DikA
i
D jkA
j =1
i
c
2 ( m 1)
,1 i c;1 k N .
(14)
47
PATTERNS 2012 : The Fourth International Conferences on Pervasive Patterns and Applications
i
1+
i = K
k =1
m
ik
2
DikA
i
t
k =1
x i = Argmax( x ij vi
x ij Bi
and Li = max(
x ij vi
i
x j Bi
A i
Bi
di =
(17)
),
(18)
A i
A i
(19)
Lni
, i = 1,, c.
(21)
j =1
x ij vi
A i
Bi
(22)
(23)
di
(16)
m
ik
where i
i _ initial =
IV.
COMPUTER SIMULATIONS
12.0
Each cluster has 200 points, while X550 is an augmented
version of X400 with an additional 150 points uniformly
distributed over [ 0,15] [ 0,11] . For data set X400 the
clustering results in Table I show that the terminal centroids
learned by all five algorithms produce good centroids.
i
a
Lni
det( Fi ))1/ n Fi 1 ( x i - vi ) = 1.
(20)
ISBN: 978-1-61208-221-9
48
PATTERNS 2012 : The Fourth International Conferences on Pervasive Patterns and Applications
x1 [ 0,100]
and
Normal( 0, 25 )
is
ISBN: 978-1-61208-221-9
X400 (centroid)
FCM: m=2
FCMPCM: =2
PFCM: a=1, b=1,
m=2, =2
PFCM: a=1,
b=0.1, m=2, =2
GKA: m=2
T2GKA: m=2
x1
4.9794
4.9407
5.0017
4.9973
4.9843
4.9566
4.9800
4.9427
4.9782
4.9397
5.0048
5.0097
X550 (centroid)
x2
5.9531
12.0593
6.0094
12.0102
5.9746
12.0506
5.9558
12.0582
5.9538
12.0568
6.0239
12.0837
x1
5.5711
5.1885
5.0076
4.9968
5.3716
5.1281
5.5410
5.1804
5.1064
5.5502
5.0137
4.9743
x2
5.4143
11.6395
6.0091
12.0103
5.7308
11.6642
5.4604
11.6445
5.4443
11.4151
5.9593
12.1031
Figure 3. The T2GKA clustering results with the proper clusters volumes
for the dataset X550, Example 1
4
3
2
1
x2
3.0
3.0
0.0
The T2GKA clustered results with the proper clusters centers
for the 5 same-circle datasets are shown in Fig. 4. For the 5
same-circle datasets, the EFCM=0.0042, EFCMPCM=0.0003,
EPFCM=0.0039 (a=1, b=0.1, m=2, =2), EPFCM=12.2009 (a=1,
b=1, m=2, =2), EGKA=0.0036, and ET2GKA=0.0026. We
compared the five clustering methods EA values. Except for
the EFCMPCM, the ET2GKA value is smaller than that in other
methods. The clustering results show the robustness of the
proposed T2GKA because a reasonable amount of noise data
does not affect its clustering performance.
Example 3: To verify that the proposed method can accord
the prototype dataset structure to learn the proper cluster
volumes, 2 artificial datasets named 7cluster and sinusoidal
were designed. There are 700 and 200 prototype data points
in the 7cluster and sinusoidal datasets, respectively. There
are 700 prototype data points in the 7cluster datasets
clustered into 7 clusters with different sizes and orientations.
Each cluster contains 100 prototype data points. The 7cluster
dataset is a mixture of two 2-variate distributions with
varying
deviation,
its
mean
vectors
5.0
1.0
1.0
5.0
2.0
2.0
4.5
are ,
, ,
,
,
, and
.
1.0 5.0 1.0 5.0 2.0 2.0
3.0
The prototype data points in the dataset sinusoidal are
generated by x2 = 10 4 sin( 0.001x12 )x13 + , where
0
-1
-2
-3
-4
-2
x1
Figure 4. Clustering results using 5 ellipsoids for the prototype data points
in the dataset 5samecircle, Example 2
normally
49
PATTERNS 2012 : The Fourth International Conferences on Pervasive Patterns and Applications
[4]
[5]
[6]
[7]
[8]
[9]
Figure 5. Clustering results using 7 ellipsoids for the prototype data points in
the dataset 7cluster, Example 3
[10]
[11]
[12]
[13]
[14]
[15]
Figure 6. Clustering results using 5 ellipsoids for the prototype data points in
the dataset sinusoidal, Example 3
V.
[16]
CONCLUSIONS
[17]
REFERENCES
[22]
[18]
[19]
[20]
[21]
[1]
ISBN: 978-1-61208-221-9
[23]
[24]
50