Skip to content

Clustering In Machine Learning Narender Kumar Spark By {Examples}

  • by

1. What are Clusters?

Clusters refer to groups of data points that share similar characteristics or features. In machine learning, clustering algorithms are used to identify these clusters or groups within a dataset based on the similarity or dissimilarity between data points.

A cluster can be defined as a set of data points that are close together in a feature space, where the distance between two data points is calculated based on their feature values. The goal of clustering is to identify these clusters and group similar data points together, while keeping dissimilar points separate.

2. What is Clustering in Machine Learning?

Clustering is a type of unsupervised machine learning technique that involves grouping similar data points together based on their features or characteristics. The goal of clustering is to identify patterns or structures within the data that are not immediately apparent, such as clusters, outliers, or subgroups.

In clustering, the data points are not labeled or pre-assigned to any particular category or class. Instead, the algorithm attempts to find natural groupings or clusters within the data by measuring the similarity or dissimilarity between each data point and all other points in the dataset.

There are several clustering algorithms available in machine learning, including k-means, hierarchical clustering, DBSCAN, and Gaussian mixture models. Each algorithm has its own strengths and weaknesses and is suited to different types of data and applications.

3. Why Clustering? 

Clustering is a useful technique in machine learning for a variety of reasons:

Identifying patterns: Clustering can help identify patterns and structures within complex datasets that are not immediately apparent. By grouping similar data points together, clustering can reveal underlying trends, relationships, and insights that may not be visible by just examining the raw data.

Data exploration: Clustering can be used for data exploration to gain a better understanding of the data and its properties. By visualizing the clusters, analysts can identify potential outliers, anomalies, and areas of interest in the data.

Data preprocessing: Clustering can be used as a preprocessing step to reduce noise and simplify the data before applying other machine learning techniques. By grouping similar data points together, clustering can help remove redundancy and improve the quality of the data.

Customer segmentation: Clustering can be used in marketing and customer analytics to group customers based on their purchasing behavior, demographics, and preferences. This can help businesses target their marketing campaigns and personalize their products and services to different customer segments.

Anomaly detection: Clustering can be used for anomaly detection to identify data points that do not belong to any cluster or are significantly different from other data points in a cluster. This can help detect unusual or potentially fraudulent behavior in financial transactions, network traffic, or medical data.

4. Examples of Clustering

Sure, here are some examples of clustering in points:

In a dataset of customer transactions, clustering can be used to group customers based on their purchasing behavior. For example, customers who frequently purchase items together or who have similar purchase histories can be grouped together into clusters.

In a dataset of medical records, clustering can be used to group patients based on their symptoms and medical history. This can help identify potential disease clusters or outbreaks, and help with diagnosis and treatment.

In a dataset of sensor data from a smart home, clustering can be used to group similar data points together based on their attributes. For example, data from sensors that measure temperature and humidity levels can be clustered to identify different patterns of usage or environmental conditions.

In a dataset of social media posts, clustering can be used to group posts based on their content or sentiment. This can help identify trends or topics of interest, and help with targeted advertising and marketing campaigns.

In a dataset of customer feedback surveys, clustering can be used to group feedback based on the topics or themes mentioned. This can help identify areas where improvements are needed or where the company is doing well.

5. Common Clustering Algorithms

There are many clustering algorithms available in machine learning, each with its own strengths and weaknesses. Here are some of the most commonly used clustering algorithms:

5.1 K-means:

K-means is a popular algorithm for clustering that involves partitioning the data into a predetermined number of clusters (k). The algorithm iteratively assigns data points to the nearest centroid (cluster center) based on their distance and updates the centroid until the optimal clusters are obtained.

Data: https://github.com/Narenderbeniwal/Spark-By-Example

# Import necessary modules
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
from sklearn.cluster import KMeans
from sklearn.preprocessing import StandardScaler

# Load the dataset
data = pd.read_csv(‘iris.data’,
header=None, names=[‘sepal_length’, ‘sepal_width’,
‘petal_length’, ‘petal_width’, ‘class’])

# Separate the features from the class labels
X = data.iloc[:, :-1].values

# Standardize the features
scaler = StandardScaler()
X = scaler.fit_transform(X)

# Create a KMeans object with 3 clusters
kmeans = KMeans(n_clusters=3)

# Fit the data to the KMeans model
kmeans.fit(X)

# Predict the cluster labels for each data point
labels = kmeans.predict(X)

# Add the cluster labels to the dataframe
data[‘cluster’] = labels

# Plot the clusters
fig, ax = plt.subplots(figsize=(8, 6))
plt.scatter(X[labels == 0, 0], X[labels == 0, 1], s=100, c=’red’, label=’Cluster 1′)
plt.scatter(X[labels == 1, 0], X[labels == 1, 1], s=100, c=’blue’, label=’Cluster 2′)
plt.scatter(X[labels == 2, 0], X[labels == 2, 1], s=100, c=’green’, label=’Cluster 3′)
plt.scatter(kmeans.cluster_centers_[:, 0], kmeans.cluster_centers_[:, 1], s=200, marker=’*’, c=’black’, label=’Centroids’)
plt.title(‘K-Means Clustering’)
plt.xlabel(‘sepal length (standardized)’)
plt.ylabel(‘sepal width (standardized)’)
plt.legend()
plt.show()

We create a KMeans object with 3 clusters and fit the data to the model using the fit method. We can then use the predict method to obtain the cluster labels for each data point.

5.2 Hierarchical clustering

Hierarchical clustering is a clustering algorithm that creates a hierarchy of clusters, starting with each data point as its own cluster and then merging clusters together based on their similarity. This can be done either agglomerative (bottom-up) or divisively (top-down).

import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
import scipy.cluster.hierarchy as shc
from sklearn.preprocessing import StandardScaler

# load the dataset
data = pd.read_csv(‘https://archive.ics.uci.edu/ml/machine-learning-databases/iris/iris.data’,
header=None, names=[‘sepal_length’, ‘sepal_width’, ‘petal_length’, ‘petal_width’, ‘class’])

# separate the features from the class labels
X = data.iloc[:, :-1].values

# standardize the features
scaler = StandardScaler()
X = scaler.fit_transform(X)

# create a dendrogram to determine the number of clusters
plt.figure(figsize=(10, 7))
plt.title(“Iris Dendograms”)
dend = shc.dendrogram(shc.linkage(X, method=’ward’))

# create a Hierarchical Clustering model with 3 clusters
from sklearn.cluster import AgglomerativeClustering
hc = AgglomerativeClustering(n_clusters=3, affinity = ‘euclidean’, linkage=’ward’)

# fit the model to the data
labels = hc.fit_predict(X)

# add the cluster labels to the dataframe
data[‘cluster’] = labels

# plot the clusters
fig, ax = plt.subplots(figsize=(8, 6))
plt.scatter(X[labels == 0, 0], X[labels == 0, 1], s=100, c=’red’, label=’Cluster 1′)
plt.scatter(X[labels == 1, 0], X[labels == 1, 1], s=100, c=’blue’, label=’Cluster 2′)
plt.scatter(X[labels == 2, 0], X[labels == 2, 1], s=100, c=’green’, label=’Cluster 3′)
plt.title(‘Hierarchical Clustering’)
plt.xlabel(‘sepal length (standardized)’)
plt.ylabel(‘sepal width (standardized)’)
plt.legend()
plt.show()

Output

6. Applications of Clustering

Clustering has a wide range of applications in various fields. Here are some examples:

Customer Segmentation: Clustering is commonly used in marketing to group customers based on their buying behavior, demographics, and other relevant factors. This can help businesses to tailor their marketing strategies and product offerings to specific customer segments.

Image Segmentation: Clustering is used in image processing to group pixels based on their color or intensity values. This can help to identify objects or regions within an image.

Anomaly Detection: Clustering can be used to identify unusual or anomalous behavior in data. This can be useful in detecting fraud in financial transactions or in identifying network intrusions.

Document Clustering: Clustering can be used to group similar documents together based on their content. This can help in organizing large document collections and in information retrieval.

Bioinformatics: Clustering is used to group genes, proteins, and other biological data based on their similarities. This can help in identifying relationships between different biological entities and in understanding their functions.

Recommender Systems: Clustering can be used in recommender systems to group users based on their preferences and behaviors. This can help in making personalized recommendations to users.

Social Network Analysis: Clustering can be used in social network analysis to group individuals based on their social connections and interactions. This can help in identifying communities or groups within a network.

Natural Language Processing: Clustering can be used in text mining to group similar documents, words, or phrases together. This can help in identifying themes, topics, or sentiments within a large corpus of text.

Market Segmentation: Clustering can be used to group similar products or services together based on their attributes, features, or benefits. This can help in creating product portfolios or pricing strategies.

Geographic Clustering: Clustering can be used to group geographic locations based on their similarities in terms of population density, economic activity, or other factors. This can help in understanding regional development or urban planning.

Conclusion

In conclusion, clustering is a powerful technique in machine learning and data analysis that is used to group similar data points together. The objective of clustering is to identify patterns or structures within the data that may not be immediately apparent to the naked eye.

 1. What are Clusters? Clusters refer to groups of data points that share similar characteristics or features. In machine learning, clustering algorithms are used to identify these clusters or groups within a dataset based on the similarity or dissimilarity between data points. A cluster can be defined as a set of data points that are  Read More Machine Learning 

Leave a Reply

Your email address will not be published. Required fields are marked *