Author Name : Shreyansh Padarha
Email : mailto:[email protected]
This repo explores KMeans and Agglomerative Clustering effectiveness in simplifying large datasets for ML. Goals include dataset download, finding optimal clusters via Elbow and Silhouette methods, comparing clustering techniques, validating optimal clusters, tuning hyperparameters. Detailed explanations and analysis are provided.
To critically analyse the effectiveness of clustering in simplifying large datasets for machine learning, this lab aims to compare KMeans and Agglomerative Clustering algorithms and evaluate their performance against a classification algorithm.
- To Download the "Car Evaluation" Dataset from UCI Repository.
- To Find the optimal number of clusters using Elbow and Silhouette Method.
- To Compare KMeans and Agglomerative Clustering methods for clustering the instances in the above dataset.
- To Validate the optimal number of clusters found out in the previous question.
- To Find what hyperparameters are suitable in KMeans (n_clusters, max_iter, init, algorithm)
- To Find what hyperparameters are suitable in Agglomerative Clustering (n_clusters, metric, linkage)
- To Plot Hierarchical Clustering (Dendrogram) [Source]
- To Compare the better clustering algorithm with any classification algorithm, and write notes on the same.
The elbow method, is a graphical representation of finding the optimal 'k' (number of clusters) while conducting a clustering Analysis
Axises
- X Axis: Number of Clusters (k)
- Y_Axis: Variation (variability) in each cluster / WCSS (Within Cluster Sum of Squares)
Within Cluster Sum Of Squares (WCSS) Formula
Distance between
The silhouette Method is also a method to find the optimal number of clusters while aiding interpretation and validation of consistency within clusters of data.
The silhouette method computes silhouette coefficients of each point that measure how much a point is similar to its own cluster compared to other clusters.
Formula (Silhoette Coefficient)
Computing s(i) — silhouette coefficient or i’th point using below mentioned formula.
- Compute a(i): The average distance of that point with all other points in the same clusters.
- Compute b(i): The average distance of that point with all the points in the closest cluster to its cluster.
Silhoutte Analysis Plot
Example
Linkage: The method in which the distance between two datapoints is calculated and the method used for deciding which two data points to connect in Hierarchical Clustering
Dendrogram: A herarchichal tree and method that plots and depicts hierarchical Clustering.
Types of linkages:
-
Single linkage: This method tends to create long, stretched-out clusters. It can also result in clusters being merged incorrectly, leading to poor results.
-
Complete linkage: This method tends to create tight, compact clusters. It is less sensitive to outliers than single linkage and tends to perform well in practice.
-
Average linkage: This method is a compromise between single and complete linkage. It tends to create well-balanced clusters that are not too stretched out.
-
Weighted linkage: This method is similar to average linkage, but it gives more weight to larger clusters. This can help to prevent small clusters from being merged too early.
-
Centroid linkage: This method is based on the distance between cluster centroids. It tends to create balanced clusters that are not too stretched out. However, it can be sensitive to outliers.
-
Median linkage: This method is similar to centroid linkage, but it uses the median distance between points instead of the mean. It tends to be more robust to outliers than centroid linkage.
-
Ward linkage: This method minimizes the sum of squared differences within each cluster. It tends to create tight, compact clusters and is often used for hierarchical clustering.
The process and methods have been extensively explained in each code chunk, with the help of single line comments.
In case of any queries, please contact me at [email protected] .
It can be seen through the elbow method
, that the optimal nunmber of k clusters are 4
.
The Within Cluster Sum of Squares for k = 4, is aproximately 1100
.
It can be seen through the silhouette method, that the optimal nunmber of k clusters are 4
.
The silhouette coefficient for k = 4, is 0.136
.
Similar to the Elbow method and Silhouette comparer method used for my algorithm, SKlearn's KMeans also identifies 4 as the ideal number of clusters. Its evident from the plots that, at k = 3 and k = 4, most of the data is being representated with a high silhouette coefficient score.
Considering PC1 and PC2's spread of clusters, after dimensionality reduction from 6 features. It can be observed that:
- The spread of the class value (quality --> categorical variables) is heterogeneous in nature, implying too scatered to incoperate all 6 features. This hypothetically suggests, the quality of the car was adjourned in a subjectie manner, where attributes like door, weren't taken as a major contributing factor.
- For k = 4, KMeans has a better cluster allocation, and more homogeneous than Agglomerative Clustering.
- The colour coded nature of the scatter plot, also helps us understand how differently were the data points clustered.
- Eg. Graphically, Agglomerative has vertically clubbed the data points into a cluster, while KMeans has horizontally clubbed the data points.
-
Both Kmeans and Agglomerative have a different spread of clusters.
- Agglomerative Clustering has spread/clustered the data points more unevenly.
- KMeans has created clusters in a "very" proportionally distributed manner.
-
Understandablly, both Kmeans and Agglomerative Clustering are not a representative of the true clss values.
-
The distribution of class values across clusters is different for agglomerative and k-means clustering. For example, in agglomerative clustering, cluster 0 has the highest count of "unacc" class values, while in k-means clustering, cluster 2 has the highest count of "unacc" class values.
-
In botb Agglomerative and KMeans Cluster 1 and 3 have 0 instances of vgood quality cars.
In general, it is difficult to compare the performance of agglomerative and k-means clustering based on the given count of clusters vs Class values. However, we can say that both methods have produced different cluster distributions and class value counts, which may have implications for further analysis and decision making.
GridSearchCV
Upon using GridSearchCV, the hypertuning parameters obtained were inept. Values:
Best parameters: {'init': 'random', 'max_iter': 250, 'n_clusters': 10}
Best score: -204.5976727420548
These values are inaccurate, as the GridSearchCV is deploying an accuracy Score, from the sklearn.metrics library, for a clustering problem.
Manual Hyperparameter tuning
The method used for judging the optimum hyperparameters in Kmeans is silhouette score.
It can be seen post a thorough hyperparameter comparative analysis that, for the given/selected dataset:
- k (Number of Clusters)
- As K (number of clusters) increase, the silhouette score increses, resulting in better clustering
- At k = 4, we achieve the highest clustering score, of 0.137245.
- n_init
- As n_init, increases the silhoette score tends to increase. This makes sense, when compared to the definition of n_init
- Its the number of times the k-means algorithm is run with different centroid seeds. The final results is the best output of n_init consecutive runs in terms of inertia.
- max_iter
- The maximum number of iterations don't matter in this dataset, the silhouette score is similar amongst all different permutations.
- The highly likely reason for the same is, that for this dataset the centroids stop changing after few iterations, resulting in the algorithm assigning same clusters to the data point till the maximum iterations are achieved.
- init (initialisation method)
- Both random and kmeans++, give similar silhouette scores, and clusters. Random is a bit more over the place, while k_means is less spread out, and more linear in fashion. Looking at the difference between them theoretically can help us better understand the observation:
- ‘k-means++’ : selects initial cluster centroids using sampling based on an empirical probability distribution of the points’ contribution to the overall inertia.
- ‘random’: choose n_clusters observations (rows) at random from data for the initial centroids.
Optimum set of parameters for KMeans Clustering Algorithm
k | init(method) | max_iterations | n_init | silhouette_score |
---|---|---|---|---|
4 | k-means++ | 100 | 5 | 0.137245 |
4 | k-means++ | 100 | 10 | 0.137245 |
4 | k-means++ | 500 | 5 | 0.137245 |
4 | k-means++ | 500 | 10 | 0.137245 |
4 | k-means++ | 1000 | 5 | 0.137245 |
4 | k-means++ | 1000 | 10 | 0.137245 |
Suggested parameters
As seen from the analysis that for the given dataset the init values and iterations don't contribute significantly. Therefore, we will chose the lower limit of these values in order to decrease computation time and increase efficiency.
n_clusters = 4
init = 'k-means++'
max_iter = 100
n_init = 5
The method used for judging the optimum hyperparameters in Agglomerative clustering is silhouette score.
The main three hyperparameters in Agglomerative clustering are n_clusters
, affinity (distancing measure)
, linkage type
. The rest are related to reducing computation time and increasing efficiency.
These are the hyperparametric observations for the selected datast:
- K (number of clusters)
- Generally, at k=2, the silhouette score was extremely hig when compared to other k values.
- Silhouette Score was higher at 2,3,4 than any other k value
- Silhouette score gradually decreased from 2 to 5 for ward and average linkage, then plateauing.
- Silhouette scored decreased minimalistically from 2 to 4, for __complete __linkage, then plateauing.
- Silhouette scored decreased rapidly from 2 to 6, almost dopping to zero for single linkage then increasing gradually till k = 10
- Affinity (distancing measure)
- Silhouette score was similar for l1, l2 and manhattan distancing measures (affinity)
- Silhouette score was slightly better for euclidean distancing measure.
- Linkage Type
- Ward Linkage is significantly better than all other measures.
- Average linkage has the second best average silhouette score amongs the linkage methods.
- Single linkage method, has the worst silhouette.
Interesting Notes
a) ward method only works with euclidean method of distancing
b) cosine affinity cant be applied for datasets which have 0, as a value.
Suggested parameters (highest silhouette score)
k (number of clusters) = 2, linkage = "ward", affinity = "euclidean"
The plots, support the the definitions of the types of linkages in the theory section. It is evident that different linkage methods provide different hierarchical structures. Its also interesting to note that, different linkage methods provide different optimum clusters to be taken. For the 25 sample data points taken from the dataset:
- Ward,Complete,Average linkage dendrograms provide a clean, intricate segregation of clusters.
- Median, Single, Centroid produce dendrograms have complicated and further apart linkages.
Although not directly comparable, the accuracy or more fitting model for the given dataset will be the Decision tree. The accuracy score between 0,1 came out to be 0,827 for the decision tree which is relatively very high. On the other hand the silhouette coefficient score for the KMeans algo was 0.137 (between -1,1).
To give a broader perspective, my understanding post using both the algorithms extensively on the given dataset is: Interpretability
- Decision Tree models are often more interpretable than K-Means models since they can provide a clear decision-making process based on the features of the input data.
- K-Means provides cluster assignments without any clear interpretation.
Scalability
- K-Means is typically faster and more scalable than Decision Tree, especially for large datasets. Therefore, if you have a large dataset, K-Means may be a better option.
Data Characteristics
- K-Means may not work well with high-dimensional or sparse data
- Decision Tree may not be robust to noisy data.
Use case
Finally, the use case and problem you are trying to solve should also be considered. If you want to group similar data points together, K-Means may be a better option. If you want to predict the class of a new data point, Decision Tree may be a better option.
Question No. | Description | Status | |
---|---|---|---|
Q1) | Downloading the "Car Evaluation" Dataset | Completed | |
Q2) | Find optimal number of clusters (Elbow & Silhouette Mehod) | Completed | |
Q3) | Compare & Validate KMeans and Agglomerative Clustering methods | Completed | |
Q4) | Hyperparameter Tuning (Manually) for Kmeans | Completed | |
Q5) | Hyperparameter Tuning (Manually) for _Agglomerative Clustering | Completed | |
Q6) | Plotting Hierarchical Clustering (Dendrogram) | Completed | |
Q7) | Comparing the clustering algorithms with any classification algorithm | Completed |
- Using both the Elbow and Silhouette methods, the optimal number of clusters for the dataset used is four.
- The Within Cluster Sum of Squares for k = 4 is approximately 1100.
- The Silhouette Score Comparison using K-means algorithm and SKlearn library's KMeans both suggest that k=4 is the ideal number of clusters.
- KMeans clustering has better cluster allocation and is more homogeneous than Agglomerative Clustering for k=4.
- Both KMeans and Agglomerative Clustering have a different spread of clusters. Agglomerative Clustering has spread/clustered the data points more unevenly, while KMeans has created clusters in a "very" proportionally distributed manner.
- The distribution of class values across clusters is different for Agglomerative and KMeans clustering.
- Hyperparameter tuning for KMeans clustering using GridSearchCV gave inept results.
- Silhouette score was used to judge the optimum hyperparameters in KMeans.
- As K (number of clusters) increases, the silhouette score increases, resulting in better clustering. At k=4, the highest clustering score of 0.137245 is achieved.
- The optimum set of parameters for KMeans clustering algorithm are k=4, init (method)=k-means++, max_iterations=100, n_init=5 with a silhouette score of 0.137245.
Overall, these observations suggest that KMeans clustering performs better than Agglomerative Clustering for the given dataset, and the optimal number of clusters is four. It is also recommended to use Silhouette score as a means of judging the optimum hyperparameters in KMeans clustering.
This lab gave me valuable insights into how unsupervised learning can be utilised for classifying datasets into clusters without labels. There were also the usual, bugging-debugging conundrums which made my logical reasoning stronger. It also made me realise that, understanding an Unsupervised learning's accuracy or precision is extremely difficult, as there are no target values to compare with. In such a situation, knowledge about the domain, one is working in becomes critical.
Car Evaluation Dataset: https://archive.ics.uci.edu/ml/datasets/Car+Evaluation
- Stackoverflow: https://stackoverflow.com/
- SKlearn Documentation: https://scikit-learn.org/stable/
- Yellowbrick Documentation: https://www.scikit-yb.org/en/latest/
- Seaborn Documentation: https://seaborn.pydata.org/
- Pandas Documentation: https://pandas.pydata.org/docs/
- matplotlib Documentation: https://matplotlib.org/stable/index.html
- Medium Blogs: https://medium.com/