applied soft computing - inaoeariel/k-harmonic means type... · clustering categorical cal...

11
Applied Soft Computing 48 (2016) 39–49 Contents lists available at ScienceDirect Applied Soft Computing journal homepage: www.elsevier.com/locate/asoc K-Harmonic means type clustering algorithm for mixed datasets Amir Ahmad ,1 , Sarosh Hashmi Faculty of Computing and Information Technology, King Abdulaziz University, Rabigh, Saudi Arabia a r t i c l e i n f o Article history: Received 14 December 2015 Received in revised form 29 February 2016 Accepted 16 June 2016 Available online 29 June 2016 Keywords: Clustering Categorical attributes Numeric attributes Mixed data K-Harmonic means clustering a b s t r a c t K-means type clustering algorithms for mixed data that consists of numeric and categorical attributes suffer from cluster center initialization problem. The final clustering results depend upon the initial clus- ter centers. Random cluster center initialization is a popular initialization technique. However, clustering results are not consistent with different cluster center initializations. K-Harmonic means clustering algo- rithm tries to overcome this problem for pure numeric data. In this paper, we extend the K-Harmonic means clustering algorithm for mixed datasets. We propose a definition for a cluster center and a distance measure. These cluster centers and the distance measure are used with the cost function of K-Harmonic means clustering algorithm in the proposed algorithm. Experiments were carried out with pure categori- cal datasets and mixed datasets. Results suggest that the proposed clustering algorithm is quite insensitive to the cluster center initialization problem. Comparative studies with other clustering algorithms show that the proposed algorithm produce better clustering results. © 2016 Elsevier B.V. All rights reserved. 1. Introduction Partitioning of data points into different set of groups on the basis of similarity between data points is called clustering [1]. These groups are called clusters. K-means clustering method [2] is a use- ful clustering method for pure numeric datasets because of its low time complexity. The selection of initial cluster centers is a critical step in this algorithm. The final clusters are strongly related with the initial cluster centers. Selecting initial cluster centers randomly, is a popular method for this purpose. However, stable results are not achieved in different runs because different initial cluster cen- ters are generated in different runs due to random initialization of cluster centers. Several cluster center initialization methods have been suggested. These methods try to compute initial cluster cen- ters [3–6]. With these initial cluster centers, K-means clustering algorithms give stable results. K-Harmonic means (KHM) clustering algorithm tries to overcome this problem in a different way by using a different cost function [7]. Experimental results suggest that clus- tering results achieved from KHM clustering algorithm are more stable as compared to K-means clustering algorithm with random initialization of cluster centers [7]. K-means clustering algorithm and KHM clustering algorithm can handle only numeric attributes. However, some datasets have Corresponding author. E-mail address: [email protected] (A. Ahmad). 1 College of Information Technology, United Arab Emirates University, Al Ain, UAE. both categorical and numeric attributes. Many K-means type clustering methods have been proposed that extend K-means clustering algorithm for mixed datasets [8–11]. However, all these clustering algorithms suffer from cluster center initialization problem. Cluster center initialization problem has not got much attention for mixed datasets. Ji et al. [12,13] propose two cluster center initialization methods for K-means type clustering algo- rithms for mixed datasets. However, the time complexity of these methods is quadratic with respect to the number of data points; hence they are not suitable for K-means type clustering algorithms which have linear complexity with respect to the number of data points. In this paper, we suggest a clustering method that can handle mixed data and is robust results to selection of initial cluster cen- ters. Ahmad and Dey [10] propose a distance measure and a center definition for mixed datasets. Using this distance measure and a cluster center definition they propose a clustering algorithm for mixed datasets. This distance measure has also been applied for fuzzy clustering of mixed datasets [14] and subspace clustering of mixed datasets [11]. We propose that this distance measure and a new cluster center definition for the mixed datasets can be used with the cost function of KHM clustering algorithm. The proposed KHM type algorithm can handle mixed datasets and the proposed algorithm gives very stable clustering results. The paper is orga- nized in following way. The related work is presented in Section 2. Our proposed algorithm is described in Section 3. Section 4 has results and discussion. Section 5 discusses conclusion and future work. http://dx.doi.org/10.1016/j.asoc.2016.06.019 1568-4946/© 2016 Elsevier B.V. All rights reserved.

Upload: others

Post on 17-Aug-2020

10 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Applied Soft Computing - INAOEariel/K-Harmonic means type... · Clustering Categorical cal attributes Numeric to attributes Mixed data K-Harmonic means clustering ... means value

K

AF

a

ARRAA

KCCNMK

1

bgftstintcbtaaatsi

c

h1

Applied Soft Computing 48 (2016) 39–49

Contents lists available at ScienceDirect

Applied Soft Computing

journa l homepage: www.e lsev ier .com/ locate /asoc

-Harmonic means type clustering algorithm for mixed datasets

mir Ahmad ∗,1, Sarosh Hashmiaculty of Computing and Information Technology, King Abdulaziz University, Rabigh, Saudi Arabia

r t i c l e i n f o

rticle history:eceived 14 December 2015eceived in revised form 29 February 2016ccepted 16 June 2016vailable online 29 June 2016

eywords:

a b s t r a c t

K-means type clustering algorithms for mixed data that consists of numeric and categorical attributessuffer from cluster center initialization problem. The final clustering results depend upon the initial clus-ter centers. Random cluster center initialization is a popular initialization technique. However, clusteringresults are not consistent with different cluster center initializations. K-Harmonic means clustering algo-rithm tries to overcome this problem for pure numeric data. In this paper, we extend the K-Harmonicmeans clustering algorithm for mixed datasets. We propose a definition for a cluster center and a distance

lusteringategorical attributesumeric attributesixed data

-Harmonic means clustering

measure. These cluster centers and the distance measure are used with the cost function of K-Harmonicmeans clustering algorithm in the proposed algorithm. Experiments were carried out with pure categori-cal datasets and mixed datasets. Results suggest that the proposed clustering algorithm is quite insensitiveto the cluster center initialization problem. Comparative studies with other clustering algorithms showthat the proposed algorithm produce better clustering results.

© 2016 Elsevier B.V. All rights reserved.

. Introduction

Partitioning of data points into different set of groups on theasis of similarity between data points is called clustering [1]. Theseroups are called clusters. K-means clustering method [2] is a use-ul clustering method for pure numeric datasets because of its lowime complexity. The selection of initial cluster centers is a criticaltep in this algorithm. The final clusters are strongly related withhe initial cluster centers. Selecting initial cluster centers randomly,s a popular method for this purpose. However, stable results areot achieved in different runs because different initial cluster cen-ers are generated in different runs due to random initialization ofluster centers. Several cluster center initialization methods haveeen suggested. These methods try to compute initial cluster cen-ers [3–6]. With these initial cluster centers, K-means clusteringlgorithms give stable results. K-Harmonic means (KHM) clusteringlgorithm tries to overcome this problem in a different way by using

different cost function [7]. Experimental results suggest that clus-ering results achieved from KHM clustering algorithm are moretable as compared to K-means clustering algorithm with random

nitialization of cluster centers [7].

K-means clustering algorithm and KHM clustering algorithman handle only numeric attributes. However, some datasets have

∗ Corresponding author.E-mail address: [email protected] (A. Ahmad).

1 College of Information Technology, United Arab Emirates University, Al Ain, UAE.

ttp://dx.doi.org/10.1016/j.asoc.2016.06.019568-4946/© 2016 Elsevier B.V. All rights reserved.

both categorical and numeric attributes. Many K-means typeclustering methods have been proposed that extend K-meansclustering algorithm for mixed datasets [8–11]. However, allthese clustering algorithms suffer from cluster center initializationproblem. Cluster center initialization problem has not got muchattention for mixed datasets. Ji et al. [12,13] propose two clustercenter initialization methods for K-means type clustering algo-rithms for mixed datasets. However, the time complexity of thesemethods is quadratic with respect to the number of data points;hence they are not suitable for K-means type clustering algorithmswhich have linear complexity with respect to the number of datapoints.

In this paper, we suggest a clustering method that can handlemixed data and is robust results to selection of initial cluster cen-ters. Ahmad and Dey [10] propose a distance measure and a centerdefinition for mixed datasets. Using this distance measure and acluster center definition they propose a clustering algorithm formixed datasets. This distance measure has also been applied forfuzzy clustering of mixed datasets [14] and subspace clustering ofmixed datasets [11]. We propose that this distance measure and anew cluster center definition for the mixed datasets can be usedwith the cost function of KHM clustering algorithm. The proposedKHM type algorithm can handle mixed datasets and the proposedalgorithm gives very stable clustering results. The paper is orga-

nized in following way. The related work is presented in Section2. Our proposed algorithm is described in Section 3. Section 4 hasresults and discussion. Section 5 discusses conclusion and futurework.
Page 2: Applied Soft Computing - INAOEariel/K-Harmonic means type... · Clustering Categorical cal attributes Numeric to attributes Mixed data K-Harmonic means clustering ... means value

4 d Soft

2

rdt

aKc

wo

cpaauuti

ictcfcoectTTictmattaafmiKtitsaact

ataft[

0 A. Ahmad, S. Hashmi / Applie

. Related work

In this section, we will discuss K-means type clustering algo-ithms for numeric datasets, categorical datasets and mixedatasets. We will also discuss the cluster initialization methods forhese algorithms.

K-means algorithm [2] is a partitional clustering method. It usesn iterative methodology to cluster the data points into k clusters.-means clustering algorithms iteratively minimize the followingost function, �, for n data points;

=n∑i=1

‖di − Cj‖b

here n is the number of data points, Cj is the nearest cluster centerf data point di and b is an integer. For Euclidean distance, b = 2.

The means value of each attribute of all the data points in aluster is used to define the cluster center. Through the iterativeartitioning, K-means algorithm minimizes the cost function. Thislgorithm starts with k selected cluster centers. Data points aressigned to the clusters and then luster centers are recalculated bysing the data points in the cluster. These two steps are repeatedntil a predefined termination condition is achieved. The final clus-ers are related with initial cluster centers. Various cluster centersnitialization methods have been suggested.

Duda and Hart [15] propose a recursive method for computingnitial cluster centers. The method is to calculate the initial clusterenters from the clusters for another clustering problem. The ini-ial cluster centers for the cluster problem can be the final clusterenters for the cluster problem plus a data point that is farthestrom the nearest cluster center. This way the k-1 cluster centersan be used to create k cluster centers. The cluster center for thene-cluster problem is the mean of all data points. The results arextended to get two clusters. The procedure is repeated to get kluster centers. Bradley and Fayyad [3] suggest a cluster center ini-ialization procedure that uses joint probability density of the data.he algorithm initially selects J small random subsets of the data.he subsets are clustered by K-means clustering method. The data

s then clustered J times by using the cluster centers achieved bylustering of these J random subsets of the data. The initial clus-er centers are selected from these J clustering results which give

inimal value of the cost function. Khan and Ahmad [4] present anlgorithm to computer initial cluster centers. The algorithm useswo observations, the first that similar data points have same clus-er membership irrespective to the initial cluster centers. Secondly,n individual attribute can be helpful in providing some indicationbout initial cluster centers. Firstly, cluster centers are computedor individual attributes. It is assumed that each attribute is nor-

ally distributed. This information is used to compute the centern the direction of that attribute. The data is then clustered by using-means clustering algorithm with these centers. Resultant clus-

ers are used to get the cluster centers. Then K-means clusterings run on all the data. The process is done for all the attributes. Allhe clustering results are converted into strings. Data points withimilar strings form clusters. The centers of these clusters are useds initial cluster centers. Erisoglu et al. [6] suggest a method thatre based on selecting 2 of the m attributes that best represent thehange in the dataset. The distances between data points in thesewo attributes are used to compute initial cluster centers.

K-means clustering method can handle only numeric attributes the means is calculated for computing cluster center and the dis-ances (like Euclidean distance) between points and cluster center

re calculated. Huang [16] presents the K-modes clustering methodor pure categorical datasets. The method replaces means of clus-ers with modes. Hamming distance is used as a distance measure16]. Generally, K-modes clustering algorithm runs with random

Computing 48 (2016) 39–49

initial cluster centers. With random initialization, clusters are notconsistent in different runs. Several cluster center initializationmethods have been suggested to overcome this problem [17–22].

Huang proposes two methods for cluster center initializationof K-modes clustering algorithm [16]. The first method suggeststhat the first k distinct data points be taken as initial cluster cen-ters. In the second method, the frequencies of all attribute valuesare calculated and the most frequent values are assigned as initialcluster centers. Sun et al. [18] show that combination of Bradley andFayyad’s algorithm [3] with K-modes clustering algorithm givesmore reliable clustering results than random initial cluster cen-ters. He [19] suggests a farthest point heuristics for initialization ofcluster centers that uses frequency of attribute values.

Cao et al. [20] present a method for cluster center initializa-tion by using a density function and a distance function. The firstcluster center is calculated by this density function. The distancebetween the probable new cluster center and already computedcluster centers are calculated. These distances along with the den-sity function are used to compute other cluster centers. Khan andAhmad [21] propose an algorithm for cluster center initializationthat combines the results of multiple clustering of a dataset basedon different initial cluster centers to get initial cluster centers.Wu et al. [22] propose a cluster centers initialization approachbased on a density function. The complexity of this algorithm isquadratic. Random subsampling is suggested to reduce the com-plexity, however, because of this step consistent clustering resultsare not guaranteed.

Some of the datasets have both numeric and categoricalattributes. K-means clustering algorithm has been extended to han-dle these mixed datasets [8]. Huang [8] presents a cost function formixed datasets along with definitions of cluster centers and dis-tances. Means values for numeric attributes and mode values forcategorical attributes are used for cluster centers. However, the useof mode as cluster center for categorical attributes results in lossof information. Also, Hamming distance is considered between acluster center and a data point which is not very accurate for multi-valued categorical attributes. To overcome these problems, Ahmadand Dey [10] present a cost function and a distance measure. Anovel definition of a cluster center is also suggested to provide abetter representation of clusters [10]. Their clustering results werecompetitive to other computationally expensive clustering algo-rithm for mixed datasets [10]. Huang et al. [9] suggest a clusteringalgorithm for mixed datasets that uses attribute weights in costfunction. Ji et al. [23] present a K-prototype clustering method formixed datasets that combine the cluster definition proposed byAhmad and Dey [10] with attribute weights proposed by Huanget al. [9]. All these clustering methods use random initialization ofcluster centers. The cluster center initialization problem has notbeen addressed for these algorithms. We found only two algo-rithms [12,13] in the literature for this purpose. Ji et al. [12,13]propose algorithms for computing cluster centers for K-means typeclustering algorithms for mixed datasets. These algorithms use theconcept of nearest neighbors and density. The complexity of thesealgorithms is quadratic which makes the algorithms unsuitable forK-means type clustering algorithms which have linear time com-plexity.

Almost all the methods that overcome cluster center initializa-tion problem in K-means clustering algorithm or K-mode clusteringalgorithm compute initial cluster centers so that these clusteringalgorithms can run with these cluster centers and give single clus-tering result. However, K-Harmonic means clustering algorithm [7]for numeric attributes gives robust results with random initializa-

tion. In other word, this method is very insensitive to the of initialcluster centers. Our proposed algorithm uses the philosophy of K-
Page 3: Applied Soft Computing - INAOEariel/K-Harmonic means type... · Clustering Categorical cal attributes Numeric to attributes Mixed data K-Harmonic means clustering ... means value

d Soft

Hn

2

icapap

dott

d

12

3

((

(

i

hcgcwcaidw

2

widvti

A. Ahmad, S. Hashmi / Applie

armonic means clustering algorithm, which [7] is discussed in theext section.

.1. K-Harmonic means (KHM) clustering algorithm

Cluster centers initialization is a problem for K-means cluster-ng algorithm. To overcome this problem Zhang [7] proposed KHMlustering algorithm with a novel cost function. KHM clusteringlgorithm uses the harmonic means of the distances from dataoints to the cluster centers in its cost function. KHM clusteringlgorithm uses the following cost function defined by � for n dataoints;

=∑n

i=1

k∑kj=1‖di − Cj‖a

(2.1)

i is the ith data point, Cj is the jth cluster center. k is the numberf desired clusters. a is a positive number. The cost function useshe squared Euclidean distance (a = 2), however, Zhang [7] suggestshat generalizing to other distance functions are possible.

KHM clustering algorithm has following steps for data having nata points and k desired clusters;

. KHM algorithm starts with random cluster centers.

. The distances between each data point to all the centers arecalculated.

. The new cluster centers are calculated.

They are calculated in following way;

(i) Calculate �(di,Cl) = ||di − Cl|| (i = 1..n, l = 1..K) (�(di,Cl) is the dis-tance between ith data point, di, and jth cluster center, Cj .Distances between each data point to each cluster center arecalculated).

(ii) �i = 1(∑k

l=11/�(di,Cl)

a

)2 (a is a positive number)

iii) qi,j = �i/�(di, Cl)a+2

iv) qj = �nj=1qi,j

(v) pi,j = qi,j

/qjvi) Cj,s = �n

i=1pi,jdi,s

where di,s is the sth attribute value of ith data point, di and Cj,ss the sth attribute value of jth cluster.

Steps 2 and 3 are continued until the cost function stabilizes.One of the important points of this algorithm is that there is no

ard cluster membership (0, 1) of each data points. Each data pointan have memberships to each cluster. Experimental results sug-est that KHM clustering algorithm is very robust to initial clusterenters and it converges faster than K-means clustering algorithmhen the initialization is far from a local optimal. We will use the

ost function of KHM clustering algorithm to develop a clusteringlgorithm for mixed datasets that will be robust to the selection ofnitial cluster centers. In the proposed algorithm, we will use theistance measure proposed for mixed data [10]. In the next section,e will discuss this distance measure.

.2. The distance measure for mixed datasets [10]

Ahmad and Dey [10] propose a distance measure that worksell for mixed datasets. This distance measure has shown promis-

ng results for clustering [10,11,14] and other applications [24]. This

istance measure uses the distances between categorical attributealues. Weights of the numeric attributes are also used in this dis-ance measure. The proposed distance measure with m attributes,n which mr attributes are numeric whereas mc attributes are cat-

Computing 48 (2016) 39–49 41

egorical is presented in Eq. (2.2). It has two components; the firstcomponent computes the distance for numeric attributes and thesecond component for categorical attributes.

�(di,Cj) is the distance between data point di and the clustercenter Cj and it is computed by using Eq. (2.2).

�(di, Cj)=mr�t=1

(wt(dit

r − Cjtr))2 +

mc�t=1�

(ditc, Cjt

c)2

(2.2)

dit r is the tth numeric attribute value of data point di and dit c isthe tth categorical attribute value of data point di. Here Cj = (Cj1,Cj2,. . .,Cjm) represents the cluster centre for jth cluster. Cjt r rep-resents the mean of tth numeric attribute for jth cluster. Cjt c

represents the centre representation for tth categorical attributesfor jth cluster. � is the distance between a categorical attributevalue and a cluster center for a categorical attribute.

In the first component, wt is the weight of the tth numericattribute. The second component uses a novel cluster representa-tion and distances between categorical attribute values [10]. Oneof the most important steps in the algorithm is the calculation ofdistances between two categorical attribute values. The distancebetween two attribute values of a categorical attribute is calcu-lated by computing the co-occurrence of these attribute values withattribute values of other categorical attributes. This algorithm hasfollowing steps;

1- Discretize all the numeric attributes to make the dataset purecategorical. This step is required to compute the distances betweeneach pair of attribute values of every categorical attribute and theweight of each numeric attribute. We would like to note that whilecalculating distances the original attribute values are used.

2- For every categorical attribute, distances between every pairof attribute values are calculated. The distance between attributevalues, x and y, of categorical attribute Ai with respect to attributecategorical Aj , for a given subset W of attribute Aj values, is defined

by �ijW and computed by using following formula;

�ijW = P(W |x) + P(∼W |y) − 1 (2.3)

P(W|x) is the probability estimates of data points with attributevalue equal to x belong to a class contained in W. Probabilityestimates are computed by computing the frequencies of pairs ofattribute values and class values.

P(∼W|y) is the probability estimates of data points with attributevalue equal to y belong to a class not contained in W. Probabilityestimates are computed by computing the frequencies of pairs ofattribute values and class values.

The distance, �ij(x,y), between x and y with respect to attributeAj is given by �ij(x,y) = P(ω|x) + P(∼ω|y) − 1, where ω is the ofsubset of values of attribute Aj that maximizes the quantityP(W|x) + P(∼W|y) − 1. This distance is calculated against every otherattribute. The mean of these distances, �(x,y), is taken as the dis-tance between x and y in the dataset. The similar exercise is donefor all the pairs of values of each categorical attribute.

3- The weight of each numeric attribute is computed. This com-putation has following steps for each numeric attribute;

(i) The discretized numeric attribute is treated as a categoricalattribute. The distance between each pair of attribute valuesis calculated by the method discussed in step 2.

(ii) The means distance of all the pairs of attribute values is used asthe weight of the numeric attribute.

4- Cluster centers are computed and then the distances between

every data point and cluster centers are calculated. In the represen-tation of a cluster center, a numeric attribute is presented by themeans of member data points, whereas, a new definition was pre-sented for categorical attributes. The cluster center for cluster C for
Page 4: Applied Soft Computing - INAOEariel/K-Harmonic means type... · Clustering Categorical cal attributes Numeric to attributes Mixed data K-Harmonic means clustering ... means value

4 d Soft Computing 48 (2016) 39–49

m

w

i

C

w

tTstd

3a

wau

wiibAAiKutef

ccicr

<

�tv

w�

c

Table 1Toy data.

Data points Attribute F Attribute G pi,j (for cluster C, the jth

cluster is C)

1 � 4.0 0.102 � 8.1 0.203 � 9.0 0.30

2 A. Ahmad, S. Hashmi / Applie

categorical attributes is defined as:

1/Nc < (N1,1,c, N1,2,c, . . .,N1,p1,c), . . .

(Ni,1,c, Ni,2,c, . . .Ni,k,c, ..,Ni,pi,c). . .. . ..

(Nm,1,c, Nm,2,c, . . .,Nm,pm,c) > (2.4)

here Cluster C has Nc data points.Ni,k,c data points in cluster C, have the kth attribute value for the

th attribute. It is assumed that ith attribute has pi different values.The distance, �(X,C), between a value X and the center of cluster

in ith dimension (ith categorical attribute) is defined as:

�(X, C) = (Ni,1,c/Nc) ∗ ı(X, Ai,1) + (Ni,2,c/ Nc) ∗ ı(X, Ai,2). . .

+ (Ni,t,c/Nc) ∗ ı(X, Ai,t) + (Ni,pi,c/ Nc) ∗ ı(X, Ai,pi) (2.5)

here Ai,t is the tth attribute value of ith categorical attribute.Ahmad and Dey [10] use this distance measure to suggest a clus-

ering method for mixed datasets. The method is presented in Fig. 1.he flow chart of the method is presented in Fig. 2. In the nextection, we will present the proposed K-Harmonic means type clus-ering algorithm for mixed datasets that uses the above discussedistance measure.

. The proposed K-Harmonic means (KHM) type clusteringlgorithm for mixed datasets

We use the distance measure discussed in previous section [10]ith the cost function [7] of KHM clustering algorithm to propose

clustering algorithm for mixed datasets. The proposed algorithmses the following cost function, �, for k number of desired clusters.

=∑n

i=1

k∑kj=1�

(di, Cj

)a (3.1)

here Cj is the nearest cluster center of data point di and �(di, Cj

)s a distance measure proposed by Ahmad and Dey [10] with mod-fied centers. a is a positive number. One of the main differencesetween the proposed algorithm and the algorithm proposed byhmad and Dey [10] is the computation of the cluster centers.hmad and Dey [10] use hard cluster memberships for comput-

ng cluster centers. However, in the proposed algorithm similar toHM clustering algorithm, soft memberships of data points will besed to compute the cluster centers. For the numeric attributes,he computation will be similar as KHM clustering algorithm, how-ver, for categorical attributes the computation will be done by theollowing procedure.

In the proposed algorithm, for a categorical attribute, the clusterenter is described by the proportional distribution of categori-al values in a cluster. Using the definition of the center of clustern fuzzy clustering algorithms for categorical datasets [25,26], theenter for a cluster C (the jth cluster) for ith categorical data isepresented as:

�i,1,C ,�i,2,C , . . .�i,k,C, ..,�m,pi,C ) > (3.2)

i,k,c is related to data points which have the kth attribute value forhe ith attribute (it is assumed that ith attribute can have pi distinctalues). We can compute �i,k,c in following way.

If kth attribute value = x,

i,k,C=n∑i=1

�(x, C)(jthcluster is C) (3.3)

here �(x,C) = pi,j for data points which have ith attribute value = x,(x,C) = 0 for data points which have ith attribute value /= x.

pi,j is calculated as suggested in KHM clustering algorithm (dis-ussed in Section 2.1).

4 � 5.0 0.255 � 3.0 0.15

A toy dataset will be used to show the calculation of a clus-ter center and distance between a data point and a cluster center.There are 5 data points represented by two attributes in the dataset.Attribute F is a categorical attribute which has two attribute values� and �. The attribute G is a numeric attribute. The calculation isshown for the center of cluster C.

For attribute value �, the first attribute value of attribute F (thefirst attribute), �1,1,C is calculated in following manner,

�1,1,C = 0.10 + 0.20 + 0.25 = 0.55

Whereas For attribute value �, the second attribute value ofattribute F (the first attribute), �1,2,C is calculated in following man-ner,

�1,2,C = 0.30 + 0.15 = 0.45

For numeric attributes, Euclidean distance with the weight ofeach numeric attribute is used for computing distances betweendata points and cluster centers. But for categorical attributes, theproposed center and the distances between attribute values areused. The distance between an attribute value X and center of clus-ter C in the direction of attribute F can be calculated as (using Eq.(2.5)).

�(X, C) = (�i,1,c) ∗ ı(X, Ai,1) + (�i,2,c)∗ı(X, Ai,2). . .

+ (�i,t,c)∗ı(X, Ai,t) + (�i,pi,c)∗ı(X, Ai,pi) (3.4)

where Ai,t is the tth attribute value of ith categorical attribute.The similar calculation will be done for all the categorical

attributes. The total distance is calculated by using Eq. (2.2).For the dataset given Table 1, the distance between data point 1

(attribute value � for categorical attribute F) and center of clusterC in the direction of the categorical attribute F is calculated as:

�(�, C) = �1,1,C ∗ �(�, �) + �1,2,C ∗ �(�, �)

�(�, C) = 0.55 ∗ �(�, �) + 0.45 ∗ �(�, �)(3.5)

The distance between data point 1 (attribute value 4.0 for cat-egorical attribute G) and center of cluster C in the direction of thenumeric attribute G is calculated as:

Centre = 4.0 ∗ 0.10 + 8.1 ∗ 0.20 + 9.0 ∗ 0.30

+5.0 ∗ 0.25 + 3.0 ∗ 0.15

= 6.42

Distance=(wG(4.0 − 6.42))2

(3.6)

where wG is the significance of the attribute G.Eqs. (3.5) and (3.6) are combined by using Eq. (2.2) to calculate

the distance between data point 1 and cluster center C.KHM clustering algorithm has two steps; the computation of

cluster centers and the computation of cluster membership byusing the distances between data points and cluster centers. In ouralgorithm, we use cluster centers as discussed above and the dis-

tance measure proposed by Ahmad and Dey [10]. Remaining stepsof KHM clustering algorithm are followed as discussed in Section2.1. The proposed KHM type clustering (C) algorithm for mixed (M)datasets (D) (KHMCMD) is presented in Fig. 3.
Page 5: Applied Soft Computing - INAOEariel/K-Harmonic means type... · Clustering Categorical cal attributes Numeric to attributes Mixed data K-Harmonic means clustering ... means value

A. Ahmad, S. Hashmi / Applied Soft Computing 48 (2016) 39–49 43

Ahmad_and_Dey_Clustering_Algorithm

Input – A mixed dataset with have n number of data points described by m attributes.

Output- k clusters

1. Normalize all nu meric attrib utes .

2. Discretize all nu meric attribute.

3. For each categorical attribute

(i) Compute the distances between each pair of attribute values.

4. For each numeric attribute

(i) Consider the discretized attribute as categorical attribute and compute the distances

between each pair of attribute values.

(ii) The average of all the distances (between each pai r of att ribute val ues) is ta ken as

the weight of the numeric attribute.

Take the original data wit h normalized nu meric attributes. Assign data points to k clusters

randomly.

Repeat steps A and B

(A) Calculate cluster centers for all the clusters (suggested in Section 2.2).

(B) Each data point is assigned to its closest cluster by using the distance measure (discussed in

Section 2.2).

Until no data point changes cluster membership or a pre-defined number of runs are reached.

prop

3

pcaptbtFcfOl

4

cMddfaemccarma

end.

Fig. 1. The clustering algorithm

.1. The complexity of KHMCMD algorithm

The proposed clustering algorithm has two parts; the com-utation of distances between every pair of attribute values forategorical attributes and discretized numeric attributes, and iter-tion process in which we calculate distances between all dataoints (n data points with m attributes) to all the cluster cen-er. O(m2n + m2S3) steps are required to compute the distancesetween every pair of attribute values for each attribute where S ishe means of numbers of attribute values of categorical attributes.or each iteration, distances between each data point and eachluster center are calculated. O(nKmr + nKmcS) steps are requiredor this calculation. For u iterations, the complexity of KHMCMD is(m2n + m2S3 + un(Kmr + KmcS)). This suggests that KHMCMD has

inear computational complexity to number of data points.

. Results and discussion

The proposed initialization method was tested on three pureategorical dataset and three mixed datasets taken from UCIachine Learning Repository [27]. The information about these

atasets are presented in Table 2. In our experiments, Equal-widthiscretization [28] was used to discretize the numeric attributes

or the proposed algorithm and the algorithm proposed by Ahmadnd Dey [10]. The number of bins was set to 10. Data points ofach dataset were distributed in predefined classes. The class infor-ation was taken as ground truth. The clustering results were

ompared against the ground truth. When clusters are more thanlasses, the clusters were assigned to a class such that the clustering

ccuracy was the maximum. Zhang [7] shows that a > 2 gives goodesults. a = 4 was selected for all the experiments. We used two

easures to compare the results of clustering: Average Accuracynd Standard Deviation. Assume that a dataset contains K classes

osed by Ahmad and Dey [10].

and total n data points, let ci (i = 1,..,K) be the data points correctlyassigned to class Ci, (i = 1,..,K) then Accuracy (AC) in% is defined as:

AC =∑K

i=1cin

× 100

If an algorithm is executed T times, then Average Accuracy in% iscalculated as:

AverageAccuracy (in%) = � = 1T

T∑i=1

ACi

ACi is the Accuracy in% in ith run.And, Standard Deviation is defined as:

StandardDeviation =√

1T

T∑i=1

(ACi − �)2

The Average Accuracy in% exhibits how good the cluster resultsare against the ground truth. The higher value is desirable. The max-imum value of Average Accuracy in% can be 100. Standard Deviationshows the consistency of the clustering results in different runs. Alow value of Standard Deviation is desirable. The minimum valueof Standard Deviation is 0. As different runs of clustering algorithmstart with different initial cluster centers, the low value of StandardDeviation (SD) shows that clustering algorithm is quite robust tothe selection of initial cluster centers.

4.1. The comparative study against the clustering algorithmproposed by Ahmad and Dey [10]

The clustering algorithm proposed by Ahmad and Dey [10] hasshown good results for various pure categorical datasets and mixed

datasets. Hence, we used that clustering algorithm for the compara-tive study. We call this algorithm KMCMD (K-Means type clusteringalgorithm for mixed datasets). Our proposed clustering algorithm,KHMCMD, uses the concepts of KMCMD [10], this is another reason
Page 6: Applied Soft Computing - INAOEariel/K-Harmonic means type... · Clustering Categorical cal attributes Numeric to attributes Mixed data K-Harmonic means clustering ... means value

44 A. Ahmad, S. Hashmi / Applied Soft Computing 48 (2016) 39–49

gorith

fiEa5

Fig. 2. Flow chart of the clustering al

or taking that clustering algorithm for comparative study. Random

nitialization of cluster centers was used for both the algorithms.xperiments were carried out with the number of desired clusterss the number of classes. Experiments were also carried out with 3,, 7 number of clusters. There was no specific reason for selecting

m proposed by Ahmad and Dey [10].

these numbers. The experiments were done to study the perfor-

mance of two clustering algorithms at different number of clusters.With each clustering setting, our proposed KHMCMD algorithm andKMCMD [10] algorithm were run 10 times and Average Accuracy in%and Standard Deviation are presented.
Page 7: Applied Soft Computing - INAOEariel/K-Harmonic means type... · Clustering Categorical cal attributes Numeric to attributes Mixed data K-Harmonic means clustering ... means value

A. Ahmad, S. Hashmi / Applied Soft Computing 48 (2016) 39–49 45

KHMCMD

Input – A mixed dataset with have n number of data points described by m attributes.

Output- K clusters

5. Normalize all nu meric attrib utes .

6. Discretize all nu meric attribute.

7. For each categorical attribute

Calculate the distances between every pair of attribute values.

8. For each numeric attribute

(iii ) Consider the discretized attribute as categ oric al att ribut e and co mpute the distances

between every pair of attribute values.

(iv) The average of all the distances (between ev ery pair of attribute values) is taken as

the weight of the numeric attribute.

9. Take the or iginal dat a wit h normalized nu meric attributes. Assign data points to k clusters

randomly. Compute the center of the cluster by using the method suggested in Section 3

with hard membership.

Run following steps till the data points don’t change their clusters or a predefined nu mber of

iterations have been finished.

(i) Compute the d ista nce bet ween eac h data point to all the clu ster ce nters as discussed in

Section 3.

(ii) The nea rest cluste r ce nte r is us ed to all ocate data points to a cluster. This ste p is used

only to check whether data points are changing their cl usters or not.

(iii ) Co mpute the membership of each data point to each clu ster usin g the method

suggested in K-Harmonic mean clustering algorithm (discussed in Section 2.1).

(iv) Compute the centers as discussed in Section 3.

Fig. 3. KHMCMD clustering algorithm.

Table 2Description of the datasets used in the experiments.

Dataset Type Data points Attributes Classes Source

Categorical Numeric

Breast Cancer Categorical 699 9 0 2 [27]Lung Cancer Categorical 32 56 0 3 [27]Mushroom Categorical 8124 22 0 2 [27]Heart-1 Mixed 303 8 5 2 [27]Heart-2 Mixed 270 7 6 2 [27]Credit Mixed 690 8 6 2 [27]

Table 3Average Accuracy (in%) and Standard Deviation for Breast Cancer dataset for KHMCMDand KMCMD with varying number of clusters.

Algorithm Number ofClusters

AverageAccuracy (in%)

StandardDeviation

KHMCMD 2 96.3 0.00KMCMD 96.1 0.10KHMCMD 3 96.3 0.00KMCMD 97.5 0.06KHMCMD 5 96.3 0.12KMCMD 97.2 0.39

4

rrnar

Table 4Average Accuracy (in%) and Standard Deviation for Lung Cancer dataset for KHMCMDand KMCMD with three clusters.

Algorithm Number ofClusters

AverageAccuracy (in%)

StandardDeviation

KHMCMD 7 96.3 0.25KMCMD 96.5 0.28

.1.1. Breast cancer datasetThis is a pure categorical dataset. Missing attribute values were

eplaced by the mode of that attribute. We executed KMCMD algo-

ithm [10] and the proposed KHMCMD algorithm with varyingumber of clusters (2, 3, 5 and 7). With each cluster setting, bothlgorithms were executed ten times. Table 3 and Fig. 4 show theesults in the form of Average Accuracy and Standard Deviation. The

KHMCMD 3 61.1 0.028KMCMD 52.6 0.052

results show that Average Accuracy of KHMCMD is quite similar toKMCMD. Considering Standard Deviation, KHMCMD means alwaysexhibited a lower value. This shows that the proposed K-Harmonicclustering algorithm is producing consistent clustering results.

4.1.2. Lung cancer datasetThis is a pure categorical dataset. Missing attribute values were

replaced by the mode of the attribute. The number of clusters was

taken as three which is equal to the number of classes. As thenumber of data points was small (32), we did not do experimentswith more number of clusters. Table 4 and Fig. 5 show the Aver-age Accuracy (in%) and Standard Deviation. The results indicate that
Page 8: Applied Soft Computing - INAOEariel/K-Harmonic means type... · Clustering Categorical cal attributes Numeric to attributes Mixed data K-Harmonic means clustering ... means value

46 A. Ahmad, S. Hashmi / Applied Soft Computing 48 (2016) 39–49

0

0.1

0.2

0.3

2 clusters3 clusters5 clusters7 clusters

Standard Dev ia�o n

KHMCMD KMCMD

020406080

100

2 clusters 3 clusters 5 clusters 7 clusters

Average Acc urac y (in %)

KHMCMD KMCMD

Fig. 4. Plots for Average Accuracy (in%) and Standard Deviation for Breast Cancer dataset for KHMCMD and KMCMD with varying number of clusters.

0

0.02

0.04

0.06

3 clus ters

Standard Devia�on

KHMCMD KMCMD

020406080

100

3 clus ters

Average Accuracy (in %)

KHMCMD KMCMD

Fig. 5. Plots of Average Accuracy (in%) and Standard Deviation for Lung Cancer dataset for KHMCMD and KMCMD with three clusters.

Table 5Average Accuracy (in%) and Standard Deviation for Mushroom dataset for KHMCMDand KMCMD with varying number of clusters.

Algorithm Number ofClusters

AverageAccuracy (in%)

StandardDeviation

KHMCMD 2 89.32 0.02KMCMD 81.16 10.94KHMCMD 3 86.93 0.07KMCMD 73.86 2.47KHMCMD 5 88.91 0.08KMCMD 89.46 2.15

Ka

4

ba5cotf3SmDn

4

aaACm

Table 6Average Accuracy (in%) and Standard Deviation for Heart-1 dataset for KHMCMD andKMCMD with varying number of clusters.

Algorithm Number ofClusters

AverageAccuracy (in%)

StandardDeviation

KHMCMD 2 84.09 0.15KMCMD 83.89 0.15KHMCMD 3 80.46 0.28KMCMD 80.73 1.33KHMCMD 5 81.72 0.50KMCMD 80.79 1.77KHMCMD 7 81.85 0.57KMCMD 79.34 2.02

Table 7Average Accuracy (in%) and Standard Deviation for Heart-2 dataset for KHMCMD andKMCMD with varying number of clusters.

Algorithm Number ofClusters

AverageAccuracy (in%)

StandardDeviation

KHMCMD 2 81.63 0.33KMCMD 80.74 1.20KHMCMD 3 78.74 1.27KMCMD 78.59 3.22KHMCMD 5 80.15 1.50

KHMCMD 7 89.50 0.02KMCMD 91.38 3.06

HMCMD performed better than KMCMD due to higher accuracynd much lower standard deviation.

.1.3. Mushroom datasetThis is a pure categorical dataset. Missing values were replaced

y the mode of the attribute. We executed KMCMD and KHMCMDlgorithms on this dataset with varying number of clusters (2, 3,

and 7). With each cluster setting, both the algorithms were exe-uted ten times. Table 5 and Fig. 6 show the results in the formf Average Accuracy (in%) and Standard Deviation. The results showhat the Average Accuracy of KHMCMD is quite similar to KMCMDor two settings (5 and 7 clusters) whereas with two settings (2 and

clusters) KHMCMD performed better than KMCMD. Consideringtandard Deviation, KHMCMD exhibited superior consist perfor-ance than KMCMD by always resulting in a much lower Standardeviation. The lower Standard Deviation suggests that KHMCMD isot affected much by initial cluster centers.

.1.4. Heart-1 datasetThis is a mixed dataset. We executed KHMCMD and KMCMD

lgorithms on this dataset with varying number of clusters (2, 3, 5

nd 7). Table 6 and Fig. 7 show the results. These results show thatverage Accuracy of KHMCMD is quite similar to that of KMCMD.onsidering Standard Deviation, KHMCMD exhibited good perfor-ance than KMCMD by mostly resulting in a much lower Standard

KMCMD 77.41 2.13KHMCMD 7 80.00 1.64KMCMD 79.19 1.40

Deviation. That shows the robustness of KHMCMD for the differentinitial cluster centers.

4.1.5. Heart-2 datasetThis is a mixed dataset. We executed KHMCMD and KMCMD

algorithms on this dataset with varying number of clusters (2, 3, 5and 7). With each cluster setting, both the algorithms were exe-cuted ten times. Table 7 and Fig. 8 show the clustering results.The results show that KHMCMD exhibited similar or higher Average

Accuracy than KMCMD. Considering Standard Deviation, KHMCMDexhibited good performance than KMCMD by always resulting ina much lower Standard Deviation value except in case of 7 clus-ters where it showed a slightly higher Standard Deviation value.
Page 9: Applied Soft Computing - INAOEariel/K-Harmonic means type... · Clustering Categorical cal attributes Numeric to attributes Mixed data K-Harmonic means clustering ... means value

A. Ahmad, S. Hashmi / Applied Soft Computing 48 (2016) 39–49 47

0

5

10

2 3clusters 5clusters 7clusters clus ters

Standard Dev ia�o n

KHMCMD KMCMD

020406080

100

2 3clusters 5clusters 7clusters clus ters

Average Accuracy (in %)

KHMCMD KMCMD

Fig. 6. Plots of Average Accuracy (in%) and Standard Deviation for Mushroom dataset for KHMCMD and KMCMD with varying number of clusters.

0

0.5

1

1.5

2

2 clusters 3 clusters 5 clusters 7 clus ters

Standard Dev ia�o n

KHMCMD KMCMD

020406080

100

2 clusters 3 clusters 5 clusters 7 clus ters

Average Acc urac y (in %)

KHMCMD KMCMD

Fig. 7. Plots for Average Accuracy (in%) and Standard Deviation for Heart-1 dataset for KHMCMD and KMCMD with varying number of clusters.

0

0.5

1

1.5

2

2 clusters 3 clusters 5 clusters 7 clusters

Standard Dev ia�o n

KHMCMD KMCMD

020406080

100

2 3clusters 5clusters 7clusters clus ters

Average Accuracy (in %)

KHMCMD KMCMD

Fig. 8. Plots for Average Accuracy (in%) and Standard Deviation for Heart-2 dataset for KHMCMD and KMCMD with varying number of clusters.

Table 8Average Accuracy (in%) and Standard Deviation for Credit dataset for KHMCMD andKMCMD with varying number of clusters.

Algorithm Number ofClusters

AverageAccuracy (in%)

StandardDeviation

KHMCMD 2 85.59 0.38KMCMD 82.23 12.77KHMCMD 3 85.74 0.10KMCMD 85.07 0.54KHMCMD 5 84.86 0.62KMCMD 84.03 1.21KHMCMD 7 84.32 0.16

Ta

4

aacioet

Table 9Average Accuracy (in%) and Standard Deviation of all datasets for KHMCMD andKMCMD with clusters numbers equal to the number of classes. Bold numbers suggestbetter Average Accuracy and lower Standard Deviation.

Dataset Algorithm Number ofClusters

AverageAccuracy (in%)

StandardDeviation

Breast Cancer KHMCMD 2 96.3 0.00KMCMD 96.1 0.10

Lung Cancer KHMCMD 3 61.1 0.028KMCMD 52.6 0.052

Mushroom KHMCMD 2 89.32 0.02KMCMD 81.16 10.94

Heart-1 KHMCMD 2 84.09 0.15KMCMD 83.89 0.15

Heart-2 KHMCMD 2 81.63 0.33KMCMD 80.74 1.20

4.1.7. Summary of the results

KMCMD 84.84 0.22

he results shows the KHMCMD produce more consistent resultss compared to KMCMD.

.1.6. Credit datasetThis is a mixed dataset. We executed KHMCMD and KMCMD

lgorithms on this dataset with varying number of clusters (2, 3, 5nd 7). With each cluster setting, both the algorithms were exe-uted ten times. Table 8 and Fig. 9 show the average accuracyn% and SD values. These results show that the Average Accuracy

f KHMCMD is quite similar to KMCMD, with all settings. Consid-ring Standard Deviation, KHMCMD exhibited good performancehan KMCMD by always resulting in a lower Standard Deviation

Credit KHMCMD 2 85.59 0.38KMCMD 82.23 12.77

value. Lower values of Standard Deviation show the robustness ofKHMCMD against cluster centers initialization problem.

We present the results for all the datasets in Table 9. For bet-ter comparative study, results are presented only for the cases inwhich the number of clusters is same as the number of classes.

Page 10: Applied Soft Computing - INAOEariel/K-Harmonic means type... · Clustering Categorical cal attributes Numeric to attributes Mixed data K-Harmonic means clustering ... means value

48 A. Ahmad, S. Hashmi / Applied Soft Computing 48 (2016) 39–49

0

5

10

15

2 clus ters 3 clus ters 5 clus ters 7 clus ters

Standard Dev ia�o n

KHMCMD KMCMD

020406080

100

2 clus ters 3 clus ters 5 clus ters 7 clus ters

Average Acc urac y (in %)

KHMCMD KMCMD

Fig. 9. Plots for Average Accuracy (in%) and Standard Deviation for Credit dataset for KHMCMD and KMCMD with varying number of clusters.

Table 10Performance of the proposed algorithm against other cluster center initialization algorithms for categorical datasets. Results are presented in clustering accuracy in%. Thebold number in a row shows the best performance for the dataset.

Dataset K-Modes with RandomInitialization (Averageresults) [16]

K-Modes with Wu [22]cluster initializationmethod

K-Modes with Cao [19]cluster initializationmethod

K-Modes with Khanand Ahmad [21]initialization method.

KHMCMD (Averageresults)

Breast Cancer 83.64 91.13 91.13 91.27 96.30Lung Cancer 52.10 50.00 50.00 50.00 61.10Mushroom 72.31 87.54 87.54 88.15 88.66

Table 11The performance of the proposed KHMCMD against other mixed data clustering algorithms. (- shows results not available). The results are presented in clustering accuracyin%. The bold numbers in a row shows the best performance for the dataset.

Dataset Huang’s K-prototype algorithm[8] (Average Results)

Similarity-BasedAgglomerative Clustering [29]

Improved K-prototypeclustering algorithm [23]

Ji et al. [13] clusteringinitialization method for mixeddata clustering

KHMCMD (AverageResults)

2.6

7.9

Ttdc

4a

cacmAwpdrImett

4a

cAaw[

Heart-1 61.3 75.2 8Heart-2 60.8 – –Credit data 58.4 55.5 7

he results suggest that in these cases KHMCMD performs betterhan KMCMD in terms of higher Average Accuracy and lower Stan-ard Deviation. In other words, KHMCMD can produce accurate andonsistent clustering results.

.2. Comparison with K-modes cluster center initializationpproaches for categorical datasets

K-modes clustering algorithm is a popular algorithm used forlustering categorical datasets [16]. As discussed in Section 2, itlso suffers for cluster centers initialization problem. Various otherluster center initialization approaches have been proposed for K-odes algorithm like, Cao et al. [20], Wu et al. [22] and Khan and

hmad [21]. We compared the performance of K-modes algorithmith various cluster center initialization approaches against the

erformance of our proposed algorithm. For K-modes with ran-om initialization method and the proposed KHMCMD the averageesults of ten runs are presented. Results are shown in Table 10.t can be observed that the proposed approach outperforms other

ethods. It is noted that other than random initialization and Wut al. [22] methods, all other initialization methods presented inhe study give same clustering results in different runs, whereashe proposed method may not give unique clustering result.

.3. The comparative study with other mixed data clusteringlgorithms for mixed datasets

We also compared the clustering results with other mixed datalustering algorithm (K-prototype algorithm [8], Similarity-Based

gglomerative Clustering [29], Improved K-prototype clusteringlgorithm [23]) and Ji et al. [13] clustering initialization methodhen the number of clusters was two. Hunag’s clustering algorithm

8] was implemented and clustering results are presented for this

80.8 84.1– 81.679.4 85.6

algorithm. We ran this algorithm ten times and average results arepresented. Results for other clustering results are taken from liter-ature [13,23,29], if available in the literature. Results are presentedin Table 11. Results suggest that the proposed method performedbetter than other clustering algorithms for mixed datasets.

4.4. Discussion

As described above, we executed KHMCMD and KMCMD algo-rithms on six datasets. Three datasets were pure categoricaldatasets and three were mixed datasets. Ideally, a K-means typeclustering algorithm which is quite robust to the cluster centerinitialization problem should have high Average Accuracy and lowStandard Deviation for different runs with different initial clustercenters. It can be noted that generally KHMCMD algorithm exhibitssimilar or better accuracy against KMCMD algorithm. KHMCMDalso shows low Standard Deviation in almost all the cases. Theseresults suggest that KHMCMD is quite insensitive to cluster centerinitialization problem.

We also compared the results with other clustering algorithmsfor categorical datasets and mixed datasets. KHMCMD performedbetter than all the algorithms in term of clustering accuracy. Thisshows that KHMCMD give accurate clustering results.

5. Conclusion and future work

K-means type clustering algorithms for mixed datasets suffersfrom cluster center initialization problems as different initializationcluster centers give different clustering results. K-Harmonic means

clustering algorithm try to address the cluster center initializationproblem. K-Harmonic means clustering algorithm is quite insensi-tive to the choice of initial cluster centers [7]. However, the originalK-Harmonic means clustering algorithm works only for numeric
Page 11: Applied Soft Computing - INAOEariel/K-Harmonic means type... · Clustering Categorical cal attributes Numeric to attributes Mixed data K-Harmonic means clustering ... means value

d Soft

dtAutstpoKa

R

[

[

[

[

[

[

[

[

[

[

[

[

[

[

[

[

[

[

[

[

A. Ahmad, S. Hashmi / Applie

atasets. In this paper, we extended the K-Harmonic means clus-ering algorithm to mixed datasets. Distance measure proposed byhmad and Dey [10] and a new definition of cluster center aresed with the K-Harmonic means clustering cost function. Clus-ering results with pure categorical datasets and mixed datasetshow that the proposed clustering algorithm is quite insensitive tohe choice of initial cluster centers. The results also suggest that theroposed clustering algorithm give better clustering results againstther clustering algorithms. In future other implementations of-Harmonic means clustering [30,31] may also be examined tochieve more optimized performance.

eferences

[1] A.K. Jain, R.C. Dubes, Algorithms for Clustering Data, Prentice-Hall, Inc., 1988.[2] J. MacQueen, Some methods for classification and analysis of multivariate

observations, in: Proceedings of the Fifth Berkeley Symposium onMathematical Statistics and Probability, Volume 1: Statistics, Berkeley, Calif.,1967, pp. 281–297.

[3] P.S. Bradley, U.M. Fayyad, Refining initial points for K-means clustering,Presented at the Proceedings of the Fifteenth International Conference onMachine Learning (1998).

[4] S.S. Khan, A. Ahmad, Cluster center initialization algorithm for K-meansclustering, Pattern Recognit. Lett. 25 (2004) 1293–1302.

[5] S. Deelers, S. Auwatanamongkol, Enhancing K-means algorithm with initialcluster centers derived from data partitioning along the data axis with thehighest variance, Int. J. Comput. Sci. 2 (2007) 247–252.

[6] M. Erisoglu, N. Calis, S. Sakallioglu, A new algorithm for initial cluster centersin k-means algorithm, Pattern Recognit. Lett. 32 (2011) 1701–1705.

[7] B. Zhang, Generalized K-Harmonic Means, Hewlett-Packard LaboratorisTechnical Report, 2000.

[8] Z. Huang, Clustering large data sets with mixed numeric and categoricalvalues, Proceedings of the 1st Pacific-Asia Conference on KnowledgeDiscovery and Data Mining,(PAKDD) (1997) 21–34.

[9] J.Z. Huang, M.K. Ng, H. Rong, Z. Li, Automated variable weighting in k-meanstype clustering, IEEE Trans. Pattern Anal. Mach. Intell. 27 (2005) 657–668.

10] A. Ahmad, L. Dey, A k-mean clustering algorithm for mixed numeric andcategorical data, Data Knowl. Eng. 63 (2007) 503–527.

11] A. Ahmad, L. Dey, A k-means type clustering algorithm for subspace clusteringof mixed numeric and categorical datasets, Pattern Recognit. Lett. 32 (2011)1062–1069.

12] J. Ji, W. Pang, Y. Zheng, Z. Wang, Z. Ma, L. Zhang, A novel cluster centerinitialization method for the k-prototypes algorithms using centrality and

[

[

Computing 48 (2016) 39–49 49

distance, Appl. Math. 9 (2015) 2933–2942.13] J. Ji, W. Pang, Y. Zheng, Z. Wang, Z. Ma, An initialization method for clustering

mixed numeric and categorical data based on the density and distance, Int. J.Pattern Recognit. Artif. Intell. 29 (2015) 1550024.

14] J. Ji, W. Pang, C. Zhou, X. Han, Z. Wang, A fuzzy k-prototype clusteringalgorithm for mixed numeric and categorical data, Knowl. Based Syst. 30(2012) 129–135.

15] R.O. Duda, P.E. Hart, Pattern Classification and Scene Analysis, vol. 3, Wiley,New York, 1973.

16] Z. Huang, A fast clustering algorithm to cluster very large categorical data setsin data mining, DMKD (1997) 0.

17] Z. Huang, Extensions to the k-means algorithm for clustering large data setswith categorical values, Data Mining Knowl. Discov. 2 (1998) 283–304.

18] Y. Sun, Q. Zhu, Z. Chen, An iterative initial-points refinement algorithm forcategorical data clustering, Pattern Recognit. Lett. 23 (2002) 875–884.

19] Z. He, Farthest-point heuristic based initialization methods for k-modesclustering, arXiv preprint cs/0610043, 2006.

20] F. Cao, J. Liang, L. Bai, A new initialization method for categorical dataclustering, Expert Syst. Appl. 36 (2009) 10223–10228.

21] S.S. Khan, A. Ahmad, Cluster center initialization algorithm for K-modesclustering, Expert Syst. Appl. 40 (2013) 7444–7456.

22] S. Wu, Q. Jiang, J.Z. Huang, A new initialization method for clusteringcategorical data, in: Advances in Knowledge Discovery and Data Mining,Springer, 2007, pp. 972–980.

23] J. Ji, T. Bai, C. Zhou, C. Ma, Z. Wang, An improved k-prototypes clusteringalgorithm for mixed numeric and categorical data, Neurocomputing 120(2013) 590–596.

24] M. García-Magarinos, J.A. Vilar, A framework for dissimilarity-basedpartitioning clustering of categorical time series, Data Mining Knowl. Discov.29 (2015) 466–502.

25] D.-W. Kim, K.H. Lee, D. Lee, Fuzzy clustering of categorical data using fuzzycentroids, Pattern Recognit. Lett. 25 (2004) 1263–1271.

26] A. Ahmad, L. Dey, Algorithm for fuzzy clustering of mixed data with numericand categorical attributes, in: Distributed Computing and InternetTechnology, Springer, 2005, pp. 561–572.

27] M. Lichman. UCI machine learning repository [Online]. Available: http://archive.ics.uci.edu/ml/.

28] J. Dougherty, R. Kohavi, M. Sahami, Supervised and unsuperviseddiscretization of continuous features, Machine Learning: Proceedings of theTwelfth International Conference (1995) 194–202.

29] C. Li, G. Biswas, Unsupervised learning with mixed numeric and nominal data,IEEE Trans. Knowl. Data Eng. 14 (2002) 673–690.

30] X. Wu, B. Wu, J. Sun, S. Qiu, X. Li, A hybrid fuzzy K-harmonic means clusteringalgorithm, Appl. Math. Modell. 39 (2015) 3398–3409.

31] W.-C. Yeh, C.-M. Lai, K.-H. Chang, A novel hybrid clustering approach based onK-harmonic means using robust design, Neurocomputing 173 (2015)1720–1732, Part 3.