Liang Shan Clustering Techniques and Applications to Image Segmentation.

Slides:



Advertisements
Similar presentations
Partitional Algorithms to Detect Complex Clusters
Advertisements

Cluster Analysis: Basic Concepts and Algorithms
1 CSE 980: Data Mining Lecture 16: Hierarchical Clustering.
Hierarchical Clustering, DBSCAN The EM Algorithm
Learning Trajectory Patterns by Clustering: Comparative Evaluation Group D.
Segmentácia farebného obrazu
Unsupervised learning
10/11/2001Random walks and spectral segmentation1 CSE 291 Fall 2001 Marina Meila and Jianbo Shi: Learning Segmentation by Random Walks/A Random Walks View.
Lecture 6 Image Segmentation
Image segmentation. The goals of segmentation Group together similar-looking pixels for efficiency of further processing “Bottom-up” process Unsupervised.
CS 376b Introduction to Computer Vision 04 / 08 / 2008 Instructor: Michael Eckmann.
Normalized Cuts and Image Segmentation Jianbo Shi and Jitendra Malik, Presented by: Alireza Tavakkoli.
Clustering… in General In vector space, clusters are vectors found within  of a cluster vector, with different techniques for determining the cluster.
Region Segmentation. Find sets of pixels, such that All pixels in region i satisfy some constraint of similarity.
Announcements  Project proposal is due on 03/11  Three seminars this Friday (EB 3105) Dealing with Indefinite Representations in Pattern Recognition.
4. Ad-hoc I: Hierarchical clustering
CS 376b Introduction to Computer Vision 04 / 04 / 2008 Instructor: Michael Eckmann.
Segmentation Graph-Theoretic Clustering.
Cutting complete weighted graphs Jameson Cahill Ido Heskia Math/CSC 870 Spring 2007.
Unsupervised Learning and Data Mining
Image Segmentation A Graph Theoretic Approach. Factors for Visual Grouping Similarity (gray level difference) Similarity (gray level difference) Proximity.
© University of Minnesota Data Mining for the Discovery of Ocean Climate Indices 1 CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance.
Image Segmentation Image segmentation is the operation of partitioning an image into a collection of connected sets of pixels. 1. into regions, which usually.
Image Segmentation Rob Atlas Nick Bridle Evan Radkoff.
Segmentation Techniques Luis E. Tirado PhD qualifying exam presentation Northeastern University.
Presenter : Kuang-Jui Hsu Date : 2011/5/3(Tues.).
Segmentation using eigenvectors
Segmentation using eigenvectors Papers: “Normalized Cuts and Image Segmentation”. Jianbo Shi and Jitendra Malik, IEEE, 2000 “Segmentation using eigenvectors:
Region Segmentation Readings: Chapter 10: 10.1 Additional Materials Provided K-means Clustering (text) EM Clustering (paper) Graph Partitioning (text)
Segmentation Course web page: vision.cis.udel.edu/~cv May 7, 2003  Lecture 31.
Cut-based & divisive clustering Clustering algorithms: Part 2b Pasi Fränti Speech & Image Processing Unit School of Computing University of Eastern.
Machine Learning Problems Unsupervised Learning – Clustering – Density estimation – Dimensionality Reduction Supervised Learning – Classification – Regression.
Chapter 14: SEGMENTATION BY CLUSTERING 1. 2 Outline Introduction Human Vision & Gestalt Properties Applications – Background Subtraction – Shot Boundary.
CLUSTERING. Overview Definition of Clustering Existing clustering methods Clustering examples.
Clustering What is clustering? Also called “unsupervised learning”Also called “unsupervised learning”
CS654: Digital Image Analysis
Clustering Gene Expression Data BMI/CS 576 Colin Dewey Fall 2010.
DATA MINING WITH CLUSTERING AND CLASSIFICATION Spring 2007, SJSU Benjamin Lam.
Compiled By: Raj Gaurang Tiwari Assistant Professor SRMGPC, Lucknow Unsupervised Learning.
Image Segmentation Superpixel methods Speaker: Hsuan-Yi Ko.
CS654: Digital Image Analysis Lecture 28: Advanced topics in Image Segmentation Image courtesy: IEEE, IJCV.
 In the previews parts we have seen some kind of segmentation method.  In this lecture we will see graph cut, which is a another segmentation method.
CZ5211 Topics in Computational Biology Lecture 4: Clustering Analysis for Microarray Data II Prof. Chen Yu Zong Tel:
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
Fuzzy Pattern Recognition. Overview of Pattern Recognition Pattern Recognition Procedure Feature Extraction Feature Reduction Classification (supervised)
A Tutorial on Spectral Clustering Ulrike von Luxburg Max Planck Institute for Biological Cybernetics Statistics and Computing, Dec. 2007, Vol. 17, No.
CS 2750: Machine Learning Clustering Prof. Adriana Kovashka University of Pittsburgh January 25, 2016.
Normalized Cuts and Image Segmentation Patrick Denis COSC 6121 York University Jianbo Shi and Jitendra Malik.
Image segmentation.
Clustering Machine Learning Unsupervised Learning K-means Optimization objective Random initialization Determining Number of Clusters Hierarchical Clustering.
Data Mining and Text Mining. The Standard Data Mining process.
Clustering (2) Center-based algorithms Fuzzy k-means Density-based algorithms ( DBSCAN as an example ) Evaluation of clustering results Figures and equations.
Clustering Clustering definition: Partition a given set of objects into M groups (clusters) such that the objects of each group are ‘similar’ and ‘different’
Unsupervised Learning: Clustering
Unsupervised Learning: Clustering
Semi-Supervised Clustering
Clustering CSC 600: Data Mining Class 21.
Machine Learning Clustering: K-means Supervised Learning
Segmentation by clustering: normalized cut
Region Segmentation Readings: Chapter 10: 10
K-means and Hierarchical Clustering
Segmentation Graph-Theoretic Clustering.
Grouping.
DATA MINING Introductory and Advanced Topics Part II - Clustering
Image Segmentation CS 678 Spring 2018.
Spectral Clustering Eric Xing Lecture 8, August 13, 2010
3.3 Network-Centric Community Detection
Text Categorization Berlin Chen 2003 Reference:
Clustering Techniques
“Traditional” image segmentation
Presentation transcript:

Liang Shan Clustering Techniques and Applications to Image Segmentation

Roadmap Unsupervised learning Clustering categories Clustering algorithms K-means Fuzzy c-means Kernel-based Graph-based Q&A

Unsupervised learning Definition 1 Supervised: human effort involved Unsupervised: no human effort Definition 2 Supervised: learning conditional distribution P(Y|X), X: features, Y: classes Unsupervised: learning distribution P(X), X: features Slide credit: Min Zhang Back

Clustering What is clustering?

Clustering Definition Assignment of a set of observations into subsets so that observations in the same subset are similar in some sense

Clustering Hard vs. Soft Hard: same object can only belong to single cluster Soft: same object can belong to different clusters Slide credit: Min Zhang

Clustering Hard vs. Soft Hard: same object can only belong to single cluster Soft: same object can belong to different clusters E.g. Gaussian mixture model Slide credit: Min Zhang

Clustering Flat vs. Hierarchical Flat: clusters are flat Hierarchical: clusters form a tree Agglomerative Divisive

Hierarchical clustering Agglomerative (Bottom-up) Compute all pair-wise pattern-pattern similarity coefficients Place each of n patterns into a class of its own Merge the two most similar clusters into one Replace the two clusters into the new cluster Re-compute inter-cluster similarity scores w.r.t. the new cluster Repeat the above step until there are k clusters left (k can be 1) Slide credit: Min Zhang

Hierarchical clustering Agglomerative (Bottom up)

Hierarchical clustering Agglomerative (Bottom up) 1 st iteration 1

Hierarchical clustering Agglomerative (Bottom up) 2 nd iteration 12

Hierarchical clustering Agglomerative (Bottom up) 3 rd iteration 12 3

Hierarchical clustering Agglomerative (Bottom up) 4 th iteration

Hierarchical clustering Agglomerative (Bottom up) 5 th iteration

Hierarchical clustering Agglomerative (Bottom up) Finally k clusters left

Hierarchical clustering Divisive (Top-down) Start at the top with all patterns in one cluster The cluster is split using a flat clustering algorithm This procedure is applied recursively until each pattern is in its own singleton cluster

Hierarchical clustering Divisive (Top-down) Slide credit: Min Zhang

Bottom-up vs. Top-down Which one is more complex? Which one is more efficient? Which one is more accurate?

Bottom-up vs. Top-down Which one is more complex? Top-down Because a flat clustering is needed as a subroutine Which one is more efficient? Which one is more accurate?

Bottom-up vs. Top-down Which one is more complex? Which one is more efficient? Which one is more accurate?

Bottom-up vs. Top-down Which one is more complex? Which one is more efficient? Top-down For a fixed number of top levels, using an efficient flat algorithm like K-means, divisive algorithms are linear in the number of patterns and clusters Agglomerative algorithms are least quadratic Which one is more accurate?

Bottom-up vs. Top-down Which one is more complex? Which one is more efficient? Which one is more accurate?

Bottom-up vs. Top-down Which one is more complex? Which one is more efficient? Which one is more accurate? Top-down Bottom-up methods make clustering decisions based on local patterns without initially taking into account the global distribution. These early decisions cannot be undone. Top-down clustering benefits from complete information about the global distribution when making top-level partitioning decisions. Back

K-means Minimizes functional: Iterative algorithm: Initialize the codebook V with vectors randomly picked from X Assign each pattern to the nearest cluster Recalculate partition matrix Repeat the above two steps until convergence Data set: Clusters: Codebook : Partition matrix:

K-means Disadvantages Dependent on initialization

K-means Disadvantages Dependent on initialization

K-means Disadvantages Dependent on initialization

K-means Disadvantages Dependent on initialization Select random seeds with at least D min Or, run the algorithm many times

K-means Disadvantages Dependent on initialization Sensitive to outliers

K-means Disadvantages Dependent on initialization Sensitive to outliers Use K-medoids

K-means Disadvantages Dependent on initialization Sensitive to outliers (K-medoids) Can deal only with clusters with spherical symmetrical point distribution Kernel trick

K-means Disadvantages Dependent on initialization Sensitive to outliers (K-medoids) Can deal only with clusters with spherical symmetrical point distribution Deciding K

Try a couple of K Image: Henry Lin

Deciding K When k = 1, the objective function is Image: Henry Lin

Deciding K When k = 2, the objective function is Image: Henry Lin

Deciding K When k = 3, the objective function is Image: Henry Lin

Deciding K We can plot objective function values for k=1 to 6 The abrupt change at k=2 is highly suggestive of two clusters knee finding or elbow finding Note that the results are not always as clear cut as in this toy example Back Image: Henry Lin

Fuzzy C-means Soft clustering Minimize functional fuzzy partition matrix fuzzification parameter, usually set to 2 Data set: Clusters: Codebook : Partition matrix: K-means:

Fuzzy C-means Minimize subject to

Fuzzy C-means Minimize subject to How to solve this constrained optimization problem?

Fuzzy C-means Minimize subject to How to solve this constrained optimization problem? Introduce Lagrangian multipliers

Fuzzy c-means Introduce Lagrangian multipliers Iterative optimization Fix V, optimize w.r.t. U Fix U, optimize w.r.t. V

Application to image segmentation Original imagesSegmentations Homogenous intensity corrupted by 5% Gaussian noise Sinusoidal inhomogenous intensity corrupted by 5% Gaussian noise Back Image: Dao-Qiang Zhang, Song-Can Chen Accuracy = 96.02% Accuracy = 94.41%

Kernel substitution trick Kernel K-means Kernel fuzzy c-means

Kernel substitution trick Kernel fuzzy c-means Confine ourselves to Gaussian RBF kernel Introduce a penalty term containing neighborhood information Equation: Dao-Qiang Zhang, Song-Can Chen

Spatially constrained KFCM : the set of neighbors that exist in a window around : the cardinality of controls the effect of the penalty term The penalty term is minimized when Membership value for x j is large and also large at neighboring pixels Vice versa Equation: Dao-Qiang Zhang, Song-Can Chen

FCM applied to segmentation Original images FCM Accuracy = 96.02% KFCM Accuracy = 96.51% SKFCM Accuracy = % SFCM Accuracy = 99.34% Image: Dao-Qiang Zhang, Song-Can Chen Homogenous intensity corrupted by 5% Gaussian noise

FCM applied to segmentation FCM Accuracy = 94.41% KFCM Accuracy = 91.11% SKFCM Accuracy = 99.88% SFCM Accuracy = 98.41% Original images Image: Dao-Qiang Zhang, Song-Can Chen Sinusoidal inhomogenous intensity corrupted by 5% Gaussian noise

FCM applied to segmentation Original MR image corrupted by 5% Gaussian noise FCM result KFCM result SFCM resultSKFCM result Back Image: Dao-Qiang Zhang, Song-Can Chen

Graph Theory-Based Use graph theory to solve clustering problem Graph terminology Adjacency matrix Degree Volume Cuts Slide credit: Jianbo Shi

Problem with min. cuts Minimum cut criteria favors cutting small sets of isolated nodes in the graph Not surprising since the cut increases with the number of edges going across the two partitioned parts Image: Jianbo Shi and Jitendra Malik

Slide credit: Jianbo Shi

Algorithm Given an image, set up a weighted graph and set the weight on the edge connecting two nodes to be a measure of the similarity between the two nodes Solve for the eigenvectors with the second smallest eigenvalue Use the second smallest eigenvector to bipartition the graph Decide if the current partition should be subdivided and recursively repartition the segmented parts if necessary

Example (a) A noisy step image (b) eigenvector of the second smallest eigenvalue (c) resulting partition Image: Jianbo Shi and Jitendra Malik

Example (a) Point set generated by two Poisson processes (b) Partition of the point set

Example (a) Three image patches form a junction (b)-(d) Top three components of the partition Image: Jianbo Shi and Jitendra Malik

Example Components of the partition with Ncut value less than 0.04 Image: Jianbo Shi and Jitendra Malik

Example Back Image: Jianbo Shi and Jitendra Malik