Clustering Sunita Sarawagi http www it iitb ac
Clustering Sunita Sarawagi http: //www. it. iitb. ac. in/~sunita 1
Outline n What is Clustering n Similarity measures n Clustering Methods n Summary n References
What Is Good Clustering? n n n A good clustering method will produce high quality clusters with n high intra-class similarity n low inter-class similarity The quality of a clustering result depends on both the similarity measure used by the method and its implementation. The quality of a clustering method is also measured by its ability to discover some or all of the hidden patterns.
Chapter 8. Cluster Analysis n What is Cluster Analysis? n Types of Data in Cluster Analysis n A Categorization of Major Clustering Methods n Partitioning Methods n Hierarchical Methods n Density-Based Methods n Grid-Based Methods n Model-Based Clustering Methods n Outlier Analysis n Summary
Type of data in clustering analysis n Interval-scaled variables: n Binary variables: n Nominal, ordinal, and ratio variables: n Variables of mixed types n High dimensional data
Interval-valued variables n Standardize data n Calculate the mean absolute deviation: where n n Calculate the standardized measurement (z-score) Using mean absolute deviation is more robust than using standard deviation
Similarity and Dissimilarity Between Objects n n Distances are normally used to measure the similarity or dissimilarity between two data objects Some popular ones include: Minkowski distance: where i = (xi 1, xi 2, …, xip) and j = (xj 1, xj 2, …, xjp) are two p-dimensional data objects, and q is a positive integer n If q = 1, d is Manhattan distance
Similarity and Dissimilarity Between Objects (Cont. ) n If q = 2, d is Euclidean distance: n Properties n n n d(i, j) 0 d(i, i) = 0 d(i, j) = d(j, i) d(i, j) d(i, k) + d(k, j) Also one can use weighted distance, parametric Pearson product moment correlation, or other disimilarity measures.
Binary Variables n A contingency table for binary data Object j Object i n Simple matching coefficient (invariant, if the binary variable is symmetric): n Jaccard coefficient (noninvariant if the binary variable is asymmetric):
Dissimilarity between Binary Variables n Example n n n gender is a symmetric attribute the remaining attributes are asymmetric binary let the values Y and P be set to 1, and the value N be set to 0
Nominal Variables n n A generalization of the binary variable in that it can take more than 2 states, e. g. , red, yellow, blue, green Method 1: Simple matching n n m: # of matches, p: total # of variables Method 2: use a large number of binary variables n creating a new binary variable for each of the M nominal states
Ordinal Variables n An ordinal variable can be discrete or continuous n order is important, e. g. , rank n Can be treated like interval-scaled n n n replacing xif by their rank map the range of each variable onto [0, 1] by replacing i-th object in the f-th variable by compute the dissimilarity using methods for intervalscaled variables
Variables of Mixed Types n n A database may contain all the six types of variables n symmetric binary, asymmetric binary, nominal, ordinal, interval and ratio. One may use a weighted formula to combine their effects. n n n f is binary or nominal: dij(f) = 0 if xif = xjf , or dij(f) = 1 o. w. f is interval-based: use the normalized distance f is ordinal or ratio-scaled n compute ranks rif and n and treat zif as interval-scaled
Distance functions on high dimensional data n n n Example: Time series, Text, Images Euclidian measures make all points equally far Reduce number of dimensions: n n n choose subset of original features using random projections, feature selection techniques transform original features using statistical methods like Principal Component Analysis Define domain specific similarity measures: e. g. for images define features like number of objects, color histogram; for time series define shape based measures.
Clustering methods n n Hierarchical clustering n agglomerative Vs divisive n single link Vs complete link Partitional clustering n distance-based: K-means n model-based: EM n density-based:
Agglomerative Hierarchical clustering n n Given: matrix of similarity between every point pair Start with each point in a separate cluster and merge clusters based on some criteria: n Single link: merge two clusters such that the minimum distance between two points from the two different cluster is the least n Complete link: merge two clusters such that all points in one cluster are “close” to all points in the other.
Example agglomerative Step 0 d e Step 1 Step 2 Step 3 Step 4 de b bde a ac c Step 4 Step 3 abcde Step 2 Step 1 Step 0 divisive
A Dendrogram Shows How the Clusters are Merged Hierarchically Decompose data objects into a several levels of nested partitioning (tree of clusters), called a dendrogram. A clustering of the data objects is obtained by cutting the dendrogram at the desired level, then each connected component forms a cluster.
Partitioning Algorithms: Basic Concept n n Partitioning method: Construct a partition of a database D of n objects into a set of k clusters Given a k, find a partition of k clusters that optimizes the chosen partitioning criterion n Global optimal: exhaustively enumerate all partitions n Heuristic methods: k-means and k-medoids algorithms n k-means (Mac. Queen’ 67): Each cluster is represented by the center of the cluster n k-medoids or PAM (Partition around medoids) (Kaufman & Rousseeuw’ 87): Each cluster is represented by one of the objects in the cluster
The K-Means Clustering Method n Given k, the k-means algorithm is implemented in 4 steps: n Partition objects into k nonempty subsets n Compute seed points as the centroids of the clusters of the current partition. The centroid is the center (mean point) of the cluster. n Assign each object to the cluster with the nearest seed point. n Go back to Step 2, stop when no more new assignment.
The K-Means Clustering Method n Example
Comments on the K-Means Method n Strength n n n Relatively efficient: O(tkn), where n is # objects, k is # clusters, and t is # iterations. Normally, k, t << n. Often terminates at a local optimum. The global optimum may be found using techniques such as: deterministic annealing and genetic algorithms Weakness n Applicable only when mean is defined, then what about categorical data? n Need to specify k, the number of clusters, in advance n Unable to handle noisy data and outliers n Not suitable to discover clusters with non-convex shapes
Variations of the K-Means Method n n A few variants of the k-means which differ in n Selection of the initial k means n Dissimilarity calculations n Strategies to calculate cluster means Handling categorical data: k-modes (Huang’ 98) n Replacing means of clusters with modes n Using new dissimilarity measures to deal with categorical objects n Using a frequency-based method to update modes of clusters n A mixture of categorical and numerical data: k-prototype method
The K-Medoids Clustering Method n Find representative objects, called medoids, in clusters n PAM (Partitioning Around Medoids, 1987) n n starts from an initial set of medoids and iteratively replaces one of the medoids by one of the nonmedoids if it improves the total distance of the resulting clustering PAM works effectively for small data sets, but does not scale well for large data sets n CLARA (Kaufmann & Rousseeuw, 1990) CLARANS (Ng & Han, 1994): Randomized sampling n Focusing + spatial data structure (Ester et al. , 1995) n
Model based clustering n n Assume data generated from K probability distributions Typically Gaussian distribution Soft or probabilistic version of K-means clustering Need to find distribution parameters. EM Algorithm
EM Algorithm n n Initialize K cluster centers Iterate between two steps n n Expectation step: assign points to clusters Maximation step: estimate model parameters
Summary n n n Cluster analysis groups objects based on their similarity and has wide applications Measure of similarity can be computed for various types of data Clustering algorithms can be categorized into partitioning methods, hierarchical methods, and model-based methods Outlier detection and analysis are very useful for fraud detection, etc. and can be performed by statistical, distance-based or deviation-based approaches Acknowledgements: slides partly from Jiawei Han’s book: Data mining concepts and Techniques.
References (1) n n n n n R. Agrawal, J. Gehrke, D. Gunopulos, and P. Raghavan. Automatic subspace clustering of high dimensional data for data mining applications. SIGMOD'98 M. R. Anderberg. Cluster Analysis for Applications. Academic Press, 1973. M. Ankerst, M. Breunig, H. -P. Kriegel, and J. Sander. Optics: Ordering points to identify the clustering structure, SIGMOD’ 99. P. Arabie, L. J. Hubert, and G. De Soete. Clustering and Classification. World Scietific, 1996 M. Ester, H. -P. Kriegel, J. Sander, and X. Xu. A density-based algorithm for discovering clusters in large spatial databases. KDD'96. M. Ester, H. -P. Kriegel, and X. Xu. Knowledge discovery in large spatial databases: Focusing techniques for efficient class identification. SSD'95. D. Fisher. Knowledge acquisition via incremental conceptual clustering. Machine Learning, 2: 139 -172, 1987. D. Gibson, J. Kleinberg, and P. Raghavan. Clustering categorical data: An approach based on dynamic systems. In Proc. VLDB’ 98. S. Guha, R. Rastogi, and K. Shim. Cure: An efficient clustering algorithm for large databases. SIGMOD'98. A. K. Jain and R. C. Dubes. Algorithms for Clustering Data. Printice Hall, 1988.
References (2) n n n n n L. Kaufman and P. J. Rousseeuw. Finding Groups in Data: an Introduction to Cluster Analysis. John Wiley & Sons, 1990. E. Knorr and R. Ng. Algorithms for mining distance-based outliers in large datasets. VLDB’ 98. G. J. Mc. Lachlan and K. E. Bkasford. Mixture Models: Inference and Applications to Clustering. John Wiley and Sons, 1988. P. Michaud. Clustering techniques. Future Generation Computer systems, 13, 1997. R. Ng and J. Han. Efficient and effective clustering method for spatial data mining. VLDB'94. E. Schikuta. Grid clustering: An efficient hierarchical clustering method for very large data sets. Proc. 1996 Int. Conf. on Pattern Recognition, 101 -105. G. Sheikholeslami, S. Chatterjee, and A. Zhang. Wave. Cluster: A multi-resolution clustering approach for very large spatial databases. VLDB’ 98. W. Wang, Yang, R. Muntz, STING: A Statistical Information grid Approach to Spatial Data Mining, VLDB’ 97. T. Zhang, R. Ramakrishnan, and M. Livny. BIRCH : an efficient data clustering method for very large databases. SIGMOD'96.
- Slides: 29