1 Context-Aware Clustering Junsong Yuan and Ying Wu EECS Dept., Northwestern University.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Segmentation and Fitting Using Probabilistic Methods
Machine Learning and Data Mining Clustering
Image Indexing and Retrieval using Moment Invariants Imran Ahmad School of Computer Science University of Windsor – Canada.
Efficient Clustering of High Dimensional Data Sets with Application to Reference Matching ANSHUL VARMA FAISAL QURESHI.
EE 290A: Generalized Principal Component Analysis Lecture 6: Iterative Methods for Mixture-Model Segmentation Sastry & Yang © Spring, 2011EE 290A, University.
Unsupervised Feature Selection for Multi-Cluster Data Deng Cai et al, KDD 2010 Presenter: Yunchao Gong Dept. Computer Science, UNC Chapel Hill.
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
Learning of Pseudo-Metrics. Slide 1 Online and Batch Learning of Pseudo-Metrics Shai Shalev-Shwartz Hebrew University, Jerusalem Joint work with Yoram.
A Unified View of Kernel k-means, Spectral Clustering and Graph Cuts
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Spatial Semi- supervised Image Classification Stuart Ness G07 - Csci 8701 Final Project 1.
Lecture 4 Unsupervised Learning Clustering & Dimensionality Reduction
Recovering Articulated Object Models from 3D Range Data Dragomir Anguelov Daphne Koller Hoi-Cheung Pang Praveen Srinivasan Sebastian Thrun Computer Science.
Unsupervised Learning
Object Class Recognition using Images of Abstract Regions Yi Li, Jeff A. Bilmes, and Linda G. Shapiro Department of Computer Science and Engineering Department.
Part 3 Vector Quantization and Mixture Density Model CSE717, SPRING 2008 CUBS, Univ at Buffalo.
Clustering with Bregman Divergences Arindam Banerjee, Srujana Merugu, Inderjit S. Dhillon, Joydeep Ghosh Presented by Rohit Gupta CSci 8980: Machine Learning.
Clustering & Dimensionality Reduction 273A Intro Machine Learning.
8/10/ RBF NetworksM.W. Mak Radial Basis Function Networks 1. Introduction 2. Finding RBF Parameters 3. Decision Surface of RBF Networks 4. Comparison.
Graph-based consensus clustering for class discovery from gene expression data Zhiwen Yum, Hau-San Wong and Hongqiang Wang Bioinformatics, 2007.
SINGLE CHANNEL SPEECH MUSIC SEPARATION USING NONNEGATIVE MATRIXFACTORIZATION AND SPECTRAL MASKS Jain-De,Lee Emad M. GraisHakan Erdogan 17 th International.
CSC321: Neural Networks Lecture 12: Clustering Geoffrey Hinton.
Kaihua Zhang Lei Zhang (PolyU, Hong Kong) Ming-Hsuan Yang (UC Merced, California, U.S.A. ) Real-Time Compressive Tracking.
CSIE Dept., National Taiwan Univ., Taiwan
Generative Topographic Mapping by Deterministic Annealing Jong Youl Choi, Judy Qiu, Marlon Pierce, and Geoffrey Fox School of Informatics and Computing.
Competence Centre on Information Extraction and Image Understanding for Earth Observation 29/03/07 Blind city classification using aggregation of clusterings.
Generalized Fuzzy Clustering Model with Fuzzy C-Means Hong Jiang Computer Science and Engineering, University of South Carolina, Columbia, SC 29208, US.
Learning a Small Mixture of Trees M. Pawan Kumar Daphne Koller Aim: To efficiently learn a.
CHAPTER 7: Clustering Eick: K-Means and EM (modified Alpaydin transparencies and new transparencies added) Last updated: February 25, 2014.
Line detection Assume there is a binary image, we use F(ά,X)=0 as the parametric equation of a curve with a vector of parameters ά=[α 1, …, α m ] and X=[x.
Learning the threshold in Hierarchical Agglomerative Clustering
Low-Rank Kernel Learning with Bregman Matrix Divergences Brian Kulis, Matyas A. Sustik and Inderjit S. Dhillon Journal of Machine Learning Research 10.
Spectral Analysis based on the Adjacency Matrix of Network Data Leting Wu Fall 2009.
Bahman Bahmani Stanford University
Artificial Intelligence 8. Supervised and unsupervised learning Japan Advanced Institute of Science and Technology (JAIST) Yoshimasa Tsuruoka.
CVPR 2006 New York City Spatial Random Partition for Common Visual Pattern Discovery Junsong Yuan and Ying Wu EECS Dept. Northwestern Univ.
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
Made by: Maor Levy, Temple University Many data points, no labels 2.
Paper Reading Dalong Du Nov.27, Papers Leon Gu and Takeo Kanade. A Generative Shape Regularization Model for Robust Face Alignment. ECCV08. Yan.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
Effective Automatic Image Annotation Via A Coherent Language Model and Active Learning Rong Jin, Joyce Y. Chai Michigan State University Luo Si Carnegie.
Gaussian Mixture Models and Expectation-Maximization Algorithm.
Radial Basis Function ANN, an alternative to back propagation, uses clustering of examples in the training set.
Analysis of Bootstrapping Algorithms Seminar of Machine Learning for Text Mining UPC, 18/11/2004 Mihai Surdeanu.
LUT Method For Inverse Halftone 資工四 林丞蔚 林耿賢. Outline Introduction Methods for Halftoning LUT Inverse Halftone Tree Structured LUT Conclusion.
PART 10 Pattern Recognition 1. Fuzzy clustering 2. Fuzzy pattern recognition 3. Fuzzy image processing FUZZY SETS AND FUZZY LOGIC Theory and Applications.
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Supervised learning network G.Anuradha. Learning objectives The basic networks in supervised learning Perceptron networks better than Hebb rule Single.
DENCLUE 2.0: Fast Clustering based on Kernel Density Estimation Alexander Hinneburg Martin-Luther-University Halle-Wittenberg, Germany Hans-Henning Gabriel.
Advanced Artificial Intelligence Lecture 8: Advance machine learning.
Iterative K-Means Algorithm Based on Fisher Discriminant UNIVERSITY OF JOENSUU DEPARTMENT OF COMPUTER SCIENCE JOENSUU, FINLAND Mantao Xu to be presented.
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
Web-Mining Agents: Transfer Learning TrAdaBoost R. Möller Institute of Information Systems University of Lübeck.
Corresponding Clustering: An Approach to Cluster Multiple Related Spatial Datasets Vadeerat Rinsurongkawong and Christoph F. Eick Department of Computer.
Non-parametric Methods for Clustering Continuous and Categorical Data Steven X. Wang Dept. of Math. and Stat. York University May 13, 2010.
Intro. ANN & Fuzzy Systems Lecture 20 Clustering (1)
An unsupervised conditional random fields approach for clustering gene expression time series Chang-Tsun Li, Yinyin Yuan and Roland Wilson Bioinformatics,
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
Document Clustering with Prior Knowledge Xiang Ji et al. Document Clustering with Prior Knowledge. SIGIR 2006 Presenter: Suhan Yu.
Semi-Supervised Clustering
Jinbo Bi Joint work with Jiangwen Sun, Jin Lu, and Tingyang Xu
Latent Variables, Mixture Models and EM
AIM: Clustering the Data together
Learning with information of features
Presented by: Chang Jia As for: Pattern Recognition
Presented by Wanxue Dong
Concave Minimization for Support Vector Machine Classifiers
A Block Based MAP Segmentation for Image Compression
Presentation transcript:

1 Context-Aware Clustering Junsong Yuan and Ying Wu EECS Dept., Northwestern University

2 Contextual pattern and co-occurrences ? Spatial contexts provide useful cues for clustering

3 K-means revisit Assumption: data samples are independent Binary label indicator Limitation: contextual information of spatial dependency is not considered in clustering data samples EM Update

4 Clustering higher-level patterns Regularized k-means Distortion in original feature space Distortion in hamming space charactering contextual patterns Same as traditional K-means clustering Regularization term due to contextual patterns Not a smooth term!

5 Chicken and Egg Problem Hamming distance in clustering contextual patterns Matrix form Cannot minimize J1 and J2 separately ! J1 is coupled with J2

6 Decoupling Fix Update Fix Update

7 Nested-EM solution Nested E-step M-step Update and separately the nested-EM algorithm can converge in finite steps. Theorem of convergence

8 Simulation results (feature space)

9 Simulation results (spatial space)

10

11 K-means Initialization

12 1st round Final Phrases

13

14

15

16 Multiple-feature clustering Dataset: handwritten numerical (‘0’-‘9’) from UCI data set –Each digit has three different types of features –Contextual pattern corresponds to compositional feature Different types of features serve as contexts of each other –Clustering each type of features into 10 “words” –Clustering 10 “phrases” based on a word-lexicon of size 3x10

17 Conclusion A context-aware clustering formulation proposed –Targets on higher-level compositional patterns in terms of co-occurrences –Discovered contextual patterns can feed back to improve the primitive feature clustering An efficient nested-EM solution which is guaranteed to converge in finite steps Successful applications in image pattern discovery and multiple-feature clustering –Can be applied to other general clustering problems