Download - Cluster Analysis

Transcript
Page 1: Cluster Analysis

Cluster Analysis

Steps in conducting cluster analysis• Formulating the Problem• Selecting a Distance or Similarity Measure • Selecting a Clustering Procedure• Deciding on the Number of Clusters• Interpreting and Profiling the Clusters• Assessing Reliability and Validity

Cluster Analysis

Cluster analysis is a class of techniques used to classify objects or cases into relatively homogeneous groups called clusters. Objects in each cluster tend to be similar to each other and dissimilar to objects in the other clusters. Cluster analysis is also called classification analysis, or numerical taxonomy. Both cluster analysis and discriminant analysis are concerned with classification. However, discriminant analysis requires prior knowledge of the cluster or group membership for each object or case included, to develop the classification rule. In contrast, in cluster analysis there is no a priori information about the group or cluster membership for any of the objects. Groups or clusters are suggested by the data, not defined a priori.

Advantages of cluster analysis:

Cluster analysis is used in marketing for different purposes:

Segmenting the market: Consumers may be clustered or grouped on the basis of benefits derived from the purchase of a product. Each cluster would consist of consumers who are relatively homogeneous in terms of the benefits they seek.

Understanding the buyer behaviors: The main purpose of cluster analysis is to identify the homogeneous groups of buyers. From these groups, buying behavior can be extracted to develop a suitable marketing strategy

Identifying new product opportunities: In the competitive environment, clustering of brands and products would lead to identify potential new product opportunities.

Selecting test markets: Cities or regions can be grouped into homogeneous clusters to arrive at different marketing strategies

Reducing Data: Other multivariate technique such as multiple discriminant analysis can be applied further to cluster analysis to describe differences in consumer’s product usage behavior. It enables to reduce the data that are more manageable than individual observations.

An Ideal Clustering Situation

Page 2: Cluster Analysis

A Practical Clustering Situation

Statistics Associated with Cluster Analysis

Agglomeration schedule. An agglomeration schedule gives information on the objects or cases being combined at each stage of a hierarchical clustering process. Cluster centroid. The cluster centroid is the mean values of the variables for all the cases or objects in a particular cluster. Cluster centers. The cluster centers are the initial starting points in nonhierarchical clustering. Clusters are built around these centers, or seeds. Cluster membership. Cluster membership indicates the cluster to which each object or case belongs.

Dendrogram. A dendrogram, or tree graph, is a graphical device for displaying clustering results. Vertical lines represent clusters that are joined together. The position of the line on the scale indicates the distances at which clusters were joined. The dendrogram is read from left to right.

Variable 2

Variable 1

XXVariable 2

Variable 1

Page 3: Cluster Analysis

Distances between cluster centers. These distances indicate how separated the individual pairs of clusters are. Clusters that are widely separated are distinct, and therefore desirable. Similarity/distance coefficient matrix. A similarity/distance coefficient matrix is a lower-triangle matrix containing pair wise distances between objects or cases.

Conducting Cluster Analysis

Illustration:

An internet cafe company wants to know attitudes towards internet surfing. With the help of marketing research team, the company identified six attitude variables. Twenty respondents were asked to express their degree of with the following statements on a seven point scale (1-disagree, 7-agree). Conduct cluster analysis method using SPSS to identify homogenous customer groups based on which a suitable marketing strategy can be adopted by the company.

V1 - Internet surfing is funV2 - Surfing is bad for your budgetV3 - I combine surfing with music and gamesV4 - I try to get best information I wanted while surfingV5 - I don’t waste time in surfingV6 - You can get lot of information from various sourcesThe obtained data is as shown below and the same is given as input to SPSS software and selecting the option as cluster Analysis.

Attitudinal Data for Clustering

Formulate the Problem

Assess the Validity of Clustering

Select a Distance Measure

Select a Clustering Procedure

Decide on the Number of Clusters

Interpret and Profile Clusters

Page 4: Cluster Analysis

Case No. V1 V2 V3 V4 V5 V6

1 6 4 7 3 2 32 2 3 1 4 5 43 7 2 6 4 1 34 4 6 4 5 3 65 1 3 2 2 6 46 6 4 6 3 3 47 5 3 6 3 3 48 7 3 7 4 1 49 2 4 3 3 6 310 3 5 3 6 4 611 1 3 2 3 5 312 5 4 5 4 2 413 2 2 1 5 4 414 4 6 4 6 4 715 6 5 4 2 1 416 3 5 4 6 4 717 4 4 7 2 2 518 3 7 2 6 4 319 4 6 3 7 2 7

20 2 3 2 4 7 2

Formulate the Problem

Perhaps the most important part of formulating the clustering problem is selecting the variables on which the clustering is based.

Inclusion of even one or two irrelevant variables may distort an otherwise useful clustering solution.

Basically, the set of variables selected should describe the similarity between objects in terms that are relevant to the marketing research problem.

The variables should be selected based on past research, theory, or a consideration of the hypotheses being tested. In exploratory research, the researcher should exercise judgment and intuition.

Select a distance or similarity measure

The most commonly used measure of similarity is the Euclidean distance or its square. The Euclidean distance is the square root of the sum of the squared differences in values for each variable. Other distance measures are also available. The city-block or Manhattan distance between two objects is the sum of the absolute differences in values for each variable. The Chebychev distance between two objects is the maximum absolute difference in values for any variable.

Page 5: Cluster Analysis

If the variables are measured in vastly different units, the clustering solution will be influenced by the units of measurement. In these cases, before clustering respondents, we must standardize the data by rescaling each variable to have a mean of zero and a standard deviation of unity. It is also desirable to eliminate outliers (cases with atypical values). Use of different distance measures may lead to different clustering results. Hence, it is advisable to use different measures and compare the results.

A Classification of Clustering Procedures

Select a Clustering Procedure – Hierarchical

Hierarchical clustering is characterized by the development of a hierarchy or tree-like structure. Hierarchical methods can be agglomerative or divisive. Agglomerative clustering starts with each object in a separate cluster. Clusters are formed by grouping objects into bigger and bigger clusters. This process is continued until all objects are members of a single cluster. Divisive clustering starts with all the objects grouped in a single cluster. Clusters are divided or split until each object is in a separate cluster. Agglomerative methods are commonly used in marketing research. They consist of linkage methods, error sums of squares or variance methods, and centroid methods.

Select a Clustering Procedure – Linkage Method

SequentialThreshold

ParallelThreshold

OptimizingPartitioning

Single linkageComplete linkage

Average linkage

Clustering Procedures

NonhierarchicalHierarchical

Agglomerative Divisive

Ward’s Method

LinkageMethods

VarianceMethods

CentroidMethods

Page 6: Cluster Analysis

The single linkage method is based on minimum distance, or the nearest neighbor rule. At every stage, the distance between two clusters is the distance between their two closest points The complete linkage method is similar to single linkage, except that it is based on the maximum distance or the furthest neighbor approach. In complete linkage, the distance between two clusters is calculated as the distance between their two furthest points. The average linkage method works similarly. However, in this method, the distance between two clusters is defined as the average of the distances between all pairs of objects, where one member of the pair is from each of the clusters. Linkage Methods of Clustering

Select a Clustering Procedure – Variance Method

The variance methods attempt to generate clusters to minimize the within-cluster variance. A commonly used variance method is the Ward's procedure. For each cluster, the means for all the variables are computed. Then, for each object, the squared Euclidean distance to the cluster means is calculated . These distances are summed for all the objects. At each stage, the two clusters with the smallest increase in the overall sum of squares within cluster distances are combined. In the centroid methods, the distance between two clusters is the distance between their centroids (means for all the variables), Every time objects are grouped, a new centroid is computed. Of the hierarchical methods, average linkage and Ward's methods have been shown to perform better than the other procedures.

Single Linkage

Minimum Distance

Complete Linkage

Maximum Distance

Average Linkage

Average Distance

Cluster 1 Cluster 2

Cluster 1 Cluster 2

Cluster 1 Cluster 2

Page 7: Cluster Analysis

Other Agglomerative Clustering Methods

Select a Clustering Procedure – Nonhierarchical

The nonhierarchical clustering methods are frequently referred to as k-means clustering. These methods include sequential threshold, parallel threshold, and optimizing partitioning. In the sequential threshold method, a cluster center is selected and all objects within a prespecified threshold value from the center are grouped together. Then a new cluster center or seed is selected, and the process is repeated for the unclustered points. Once an object is clustered with a seed, it is no longer considered for clustering with subsequent seeds. The parallel threshold method operates similarly, except that several cluster centers are selected simultaneously and objects within the threshold level are grouped with the nearest center. The optimizing partitioning method differs from the two threshold procedures in that objects can later be reassigned to clusters to optimize an overall criterion, such as average within cluster distance for a given number of clusters.

Ward’s Procedure

Centroid Method

Page 8: Cluster Analysis

It has been suggested that the hierarchical and nonhierarchical methods be used in tandem. First, an initial clustering solution is obtained using a hierarchical procedure, such as average linkage or Ward's. The number of clusters and cluster centroids so obtained are used as inputs to the optimizing partitioning method. Choice of a clustering method and choice of a distance measure are interrelated. For example, squared Euclidean distances should be used with the Ward's and centroid methods. Several nonhierarchical procedures also use squared Euclidean distances.

Results of Hierarchical Clustering

Stage cluster Clusters combined first appears

Stage Cluster 1 Cluster 2 Coefficient Cluster 1 Cluster 2 Next stage1 14 16 1.000000 0 0 62 6 7 2.000000 0 0 73 2 13 3.500000 0 0 154 5 11 5.000000 0 0 115 3 8 6.500000 0 0 166 10 14 8.160000 0 1 97 6 12 10.166667 2 0 108 9 20 13.000000 0 0 119 4 10 15.583000 0 6 1210 1 6 18.500000 6 7 1311 5 9 23.000000 4 8 1512 4 19 27.750000 9 0 1713 1 17 33.100000 10 0 1414 1 15 41.333000 13 0 1615 2 5 51.833000 3 11 1816 1 3 64.500000 14 5 1917 4 18 79.667000 12 0 1818 2 4 172.662000 15 17 1919 1 2 328.600000 16 18 0

Agglomeration Schedule Using Ward’s Procedure

Page 9: Cluster Analysis

Dendrogram Using Ward’s Method

Number of Clusters

Label case 4 3 2

1 1 1 12 2 2 23 1 1 14 3 3 25 2 2 26 1 1 17 1 1 18 1 1 19 2 2 210 3 3 211 2 2 212 1 1 113 2 2 214 3 3 215 1 1 116 3 3 217 1 1 118 4 3 219 3 3 220 2 2 2

Cluster Membership of Cases Using Ward’s Procedure

Page 10: Cluster Analysis

Decide on the Number of Clusters

Theoretical, conceptual, or practical considerations may suggest a certain number of clusters. In hierarchical clustering, the distances at which clusters are combined can be used as criteria. This information can be obtained from the agglomeration schedule or from the dendrogram. In nonhierarchical clustering, the ratio of total within-group variance to between-group variance can be plotted against the number of clusters. The point at which an elbow or a sharp bend occurs indicates an appropriate number of clusters. The relative sizes of the clusters should be meaningful.

Interpreting and Profiling the Clusters

Interpreting and profiling clusters involves examining the cluster centroids. The centroids enable us to describe each cluster by assigning it a name or label. It is often helpful to profile the clusters in terms of variables that were not used for clustering. These may include demographic, psychographic, product usage, media usage, or other variables.

Cluster Centroids

The above table gives the centroid or means values for each cluster. Cluster 1 has relatively high values on the variables V1(Internet surfing is fun) and V3(I combine surfing with music and games).it has low value on V5(I don’t waste time in surfing). Hence cluster 1 can be labeled as “surf loving and concentrated”. this cluster consists of cases:1,3,6,7,,8,12,15,and 17. Cluster 2 is just opposite, with low values on V1 and V3

and a high value on V5 and this can be labeled as “apathetic surfers”. This consists of cases 2,5,9,11,13 and 20. Cluster 3 has high values on V2(Surfing is bad for your budget ), V4(I try to get best information I wanted while surfing), and V6 (You can get lot of information from various sources). Thus this cluster can be labeled as “economical surfers”. This consists of cases 4, 10,14,16,18 and 19.

Means of Variables

Cluster No. V1 V2 V3 V4 V5 V6

1 5.750 3.625 6.000 3.125 1.750 3.875

2 1.667 3.000 1.833 3.500 5.500 3.333

3 3.500 5.833 3.333 6.000 3.500 6.000

Page 11: Cluster Analysis

Assess Reliability and Validity

Perform cluster analysis on the same data using different distance measures. Compare the results across measures to determine the stability of the solutions.

Use different methods of clustering and compare the results. Split the data randomly into halves. Perform clustering separately on each half.

Compare cluster centroids across the two sub samples. Delete variables randomly. Perform clustering based on the reduced set of

variables. Compare the results with those obtained by clustering based on the entire set of variables.

In nonhierarchical clustering, the solution may depend on the order of cases in the data set. Make multiple runs using different order of cases until the solution stabilizes.

Results of Nonhierarchical Clustering

Iteration History

Convergence achieved due to no or small distance change. The maximum distance by which any center has changed is 0.000. The current iteration is 2. The minimum distance between initial centers is 7.746.

Iteration Change in clusters1 2 3

1 2.154 2.102 2.5502 0.000 0.000 0.000

Initial Cluster Centers

4 2 7

6 3 2

3 2 6

7 4 4

2 7 1

7 2 3

V1

V2

V3

V4

V5

V6

1 2 3

Cluster

Page 12: Cluster Analysis

Cluster Membership

3 1.414

2 1.323

3 2.550

1 1.404

2 1.848

3 1.225

3 1.500

3 2.121

2 1.756

1 1.143

2 1.041

3 1.581

2 2.598

1 1.404

3 2.828

1 1.624

3 2.598

1 3.555

1 2.154

2 2.102

Case Number

1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

16

17

18

19

20

Cluster Distance

Final Cluster Centers

4 2 6

6 3 4

3 2 6

6 4 3

4 6 2

6 3 4

V1

V2

V3

V4

V5

V6

1 2 3

Cluster

Distances between Final Cluster Centers

5.568 5.6985.568 6.928

5.698 6.928

Cluster

12

3

1 2 3

Page 13: Cluster Analysis

It is interesting note that clusters identified from hierarchical method are same in non-hierarchical method except change in the order. The distances between the final cluster centers indicate that the pairs of clusters are well separated.

Hierarchical clustering includes methods such as single linkage, complete linkage and average linkage. We need not to specify in advance how many clusters are to be extracted. A range of solution provided by the software from 1- cluster solution to n-cluster solution. Where as in Non-hierarchical clustering, you have to specify in advance how many clusters are required from the data. The specified number of nodes and points closest to them are used to form initial clusters and through an iterative rearrangements and the final k-clusters are determined by the package.


Top Related