Machine Learning Notes-1 (Clustering-1)
Machine Learning Notes-1 (Clustering-1)
Partitioning methods
work by iteratively
refining the cluster
centroids until
convergence is reached.
In this algorithm, we develop the hierarchy of clusters in the form of a tree, and
this tree-shaped structure is known as the dendrogram.
Agglomerative Clustering
Agglomerative is a bottom-up approach, in which the algorithm starts with
taking all data points as single clusters and merging them until one cluster is left.
Divisive Clustering
Divisive algorithm is the reverse of the agglomerative algorithm as it is a top-
down approach.
https://www.google.com/search?sca_esv=ad2aacbf6bcb86d2&q=hierarchical+clustering&tbm=isch&source=lnms&sa=X&ved=2ahUKEwiDr9vr4
dmEAxXufGwGHQdVDC0Q0pQJegQIDRAB&biw=1366&bih=587&dpr=1#imgrc=UR-6ylprlb0lAM
Density based Clustering
The basic steps to develop clustering process are presented in figure 1 and
can be summarized as follows :
Feature selection: The goal is to select properly the features on which
clustering is to be performed so as to encode as much information as
possible concerning the task of our interest.
Clustering algorithm: This step refers to the choice of an algorithm that
results in the definition of a good clustering scheme for a data set.
i) Proximity measure: It is a measure that quantifies how “similar” two
data points (i.e. feature vectors) are.
ii) Clustering criterion: In this step, we have to define the clustering
criterion, which can be expressed via a cost function or some other type
of rules. Thus, we may define a “good” clustering criterion, leading to a
partitioning that fits well the data set.
Steps of clustering process