Post on 03-Mar-2015
Clustering with Self Organizing Maps
Vahid Moosavi
Supervisor: Prof. Ludger Hovestadt
September 2011
1
Outline
• SOM Clustering Approaches
• U*C clustering (1)
– Basic Definitions
– The Algorithm
– Results
2(1):Alfred Ultsch: U*C: Self-organized Clustering with Emergent Feature Maps. LWA 2005: 240-244
The Learning Algorithm
3
Competition Cooperation And Adaptation
Representation
SOM Clustering
• One Stage Clustering: For maps with small number of nodes, each node is representative for a cluster
• Two Stage Clustering (for large maps)(1):– First train the SOM– Then apply any clustering algorithm on the nodes
instead of original data• Partitional clustering algorithms• Hierarchical clustering algorithms
• U*C Clustering Algorithm (2)
(1): VESANTO AND ALHONIEMI: CLUSTERING OF THE SELF-ORGANIZING MAP, IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 11, NO. 3, MAY 2000(2): Alfred Ultsch: U*C: Self-organized Clustering with Emergent Feature Maps. LWA 2005: 240-244 4
Weaknesses of the Existing clustering Algorithms
• Weaknesses of the other algorithms (e.g. K-Means, GK, Hierarchical Clustering):– No. of clusters should be known in advance.– Clustering algorithms are based on some geometrical
assumptions (Euclidean Distance, Ellipsoidal or spherical shapes, …)
– …
• U*C clustering improves all of the above mentioned issues.
5
U*C clustering(Basic Definitions)
• Component Planes
• U Matrix (1990)
• P Matrix (2003)
• U* Matrix
6
Presentation and visualization(Component Plane)
7
Presentation and visualizationU Matrix (1990)
8
Neuron i: ni
Neighborhood neurons of ni:N(i)
Definition of U-Matrix
Presentation and visualizationU Matrix (1990)
9
A display of all U-heights on top of the grid is called a U-Matrix: Ultsch (1990)
U-Matrix can show visually the hidden clusters in the data set
Presentation and visualizationU Matrix (1990)
10
The Original Data set
The U-Matrix
Water Shed (Border)
Basins (clusters)
Presentation and visualizationP Matrix (2003)
• In some Cases the U-Matrix is not enough. We use measure of the Density in addition to the Distance.
11
Presentation and visualizationP Matrix (2003)
• In some Cases the U-Matrix is not enough. We use measure of the Density in addition to the Distance.
12
Presentation and visualizationU* Matrix (1990)
• As the TwoDiamonds data set shows, a combination of distance relationships and density relationships is necessary to give an appropriate clustering. The combination of a U-Matrix and a P-Matrix is called U*-Matrix.
• The Main Idea: The U*-Matrix exhibits the local data distances as heights, when the data density is low (cluster border). If the data density is high, the distances are scaled down to zero (cluster center).
13
Presentation and visualizationU* Matrix (1990)
14
U*C Clustering(Main Ideas)
15
First Main Idea:Uheight in the center of a cluster is smaller than the Uheight on the border of the cluster in the U-matrix .
Second Main Idea:The P-height in the center of a cluster is larger than the P height in the border of a cluster in P-matrix.At cluster borders the local density of the points should decrease substantially
U*C Clustering(Main Ideas)
16
A movement from one position ni to another position nj with the result that wj is more within a cluster C than wi is called immersive.
•Some times, immersion can be find on U-Matrix (based on Gradient Descent Method).•Some times, immersion can be find on P-Matrix (based on Gradient Ascent Method)
•Then :1. Do Gradient Descent Method on U-Matrix:
•Start from point (Node) n in U-Matrix and go in a direction in its neighborhood to reach to minimum U-height (distance) point U. (this is probably a node within a cluster).
2. Do Gradient Ascent Method on P-Matrix:•Start from point U in P-matrix and go in a direction in its neighborhood to reach to Maximum P, Immersion Points (which will be probably the center of a Cluster)
3. Calculate the watersheds on the U*Matrix based on any existing algorithm.4. Partition Immersion Points using these water sheds to Cluster Centers C1,…,Cc.5. Assign the data sets to the clusters based on the Immersion Points of their corresponding
Unit of the SOM.
U*C Clustering (2005)
17
U*C Clustering (2005)Some Experimental Results
18
U*C Clustering (2005)Some Experimental Results
19
U*C Clustering (2005)Some Experimental Results
20
Conclusion• SOM and can transform high dimensional
Data sets to two dimensional representation and after that just by analyzing the distances and densities of the transformed data, we can find natural clusters hidden in original data sets.
21
High Dimensional Data Set
SOMModeling
Two Dimensional Representation
U-MatrixP-Matrix,…
Classification and Prediction for future
experiments
Clustering Data Sets
Conclusion• Alternative Way
22
High Dimensional Data Set
SOMModeling
Two Dimensional Representation
U-MatrixP-Matrix,…
Classification and Prediction for future
experiments
Clustering Data Sets
Feature Selection
and Extraction
Transformed (reduced) Data
set
THANKS
23