Clustering Technique: Mohammad Ali Joneidi
Clustering Technique: Mohammad Ali Joneidi
Clustering Technique: Mohammad Ali Joneidi
I. I NTRODUCTION
Clustering algorithms are generally used in an unsupervised
fashion. They are presented with a set of data instances that
must be grouped according to some notion of similarity. The
algorithm has access only to the set of features describing
each object. it is not given any information (e.g., labels) as
to where each of the instances should be placed within the
partition. However, in real application domains, it is often
the case that the experimenter possesses some background
knowledge (about the domain or the data set) that could be
useful in clustering the data. Traditional clustering algorithms
have no way to take advantage of this information even when
it does exist.In continues We use this clustering techniques
1) Hierarchy
2) K-means
3) Self Oriented Map
Fig. 2.
III. H IERARCHY
Given a set of N items (in this problem 400) to be clustered,
and an NN distance (or similarity) matrix, the basic process
of hierarchical clustering is this:
1) Start by assigning each item to its own cluster, so that
if you have N items, you now have N clusters, each
containing just one item. Let the distances (similarities)
between the clusters equal the distances (similarities)
between the items they contain.
2) Find the closest (most similar) pair of clusters and merge
them into a single cluster, so that now you have one less
cluster.
3) Compute distances (similarities) between the new cluster
and each of the old clusters.
4) Repeat steps 2 and 3 until all items are clustered into a
single cluster of size N.
Fig. 1.
V. SOM
Kohonen Self-Organizing Maps are a type of neural
network.Self- Organizing Maps are aptly named. SelfOrganizing is because no supervision is required. SOMs
learn on their own through unsupervised competitive learning.
Maps is because they attempt to map their weights to conform
to the given input data. The nodes in a SOM network attempt
to become like the inputs presented to them. In this sense,
this is how they learn. They can also be called Feature Maps,
as in Self-Organizing Feature Maps. Retaining principle
features of the input data is a fundamental principle of
SOMs, and one of the things that makes them so valuable.
The structure of a SOM is fairly simple, and is best understood
with the use of an illustration such as Figure 4
Figure 4 is a 4x4 SOM network . It is easy to overlook this
structure , but there are a few key things to notice. First, each
map node is connected to each input node. For this small
4x4 node network, that is 4x4x3=48 connections.Secondly,
notice that map nodes are not connected to each other. The
nodes are organized in this manner, as a 2-D grid makes
it easy to visualize the results. This representation is also
useful when the SOM algorithm is used. In this configuration,
each map node has a unique (i,j) coordinate. This makes it
easy to reference a node in the network, and to calculate the
distances between nodes. Because of the connections only
to the input nodes, the map nodes are oblivious as to what
values their neighbors have. A map node will only update its
weights (explained next) based on what the input vector tells it.
Fig. 6. Accuracy of Som for face detection problem for various epoch number
over fitting
VII. R EFERENCES
VI. C ONCLUSION
In hierarchy clustering we have low performance because
our data are not hierarchy base in k-means clustering with
random initial center have low accuracy for enhancing performance we consider center of each cluster one of that cluster
sample randomize.in som clustering after 2000 epoch for 400
train we have best performance and for higher epoch we have