Unsupervised-learning Methods for Image Clustering

Slides:



Advertisements
Similar presentations
Part 2: Unsupervised Learning
Advertisements

University of Joensuu Dept. of Computer Science P.O. Box 111 FIN Joensuu Tel fax Gaussian Mixture.
Image Modeling & Segmentation
Mixture Models and the EM Algorithm
Learning Representations. Maximum likelihood s r s?s? World Activity Probabilistic model of neuronal firing as a function of s Generative Model.
Supervised Learning Recap
DATA MINING van data naar informatie Ronald Westra Dep. Mathematics Maastricht University.
Machine Learning and Data Mining Clustering
Texture Segmentation Based on Voting of Blocks, Bayesian Flooding and Region Merging C. Panagiotakis (1), I. Grinias (2) and G. Tziritas (3)
Visual Recognition Tutorial
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
Paper Discussion: “Simultaneous Localization and Environmental Mapping with a Sensor Network”, Marinakis et. al. ICRA 2011.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Unsupervised Image Clustering using Probabilistic Continuous Models and Information Theoretic Principles Shiri Gordon Electrical Engineering – System,
First introduced in 1977 Lots of mathematical derivation Problem : given a set of data (data is incomplete or having missing values). Goal : assume the.
Part 4 b Forward-Backward Algorithm & Viterbi Algorithm CSE717, SPRING 2008 CUBS, Univ at Buffalo.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Visual Recognition Tutorial
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Expectation-Maximization (EM) Chapter 3 (Duda et al.) – Section 3.9
Student: Kylie Gorman Mentor: Yang Zhang COLOR-ATTRIBUTES- RELATED IMAGE RETRIEVAL.
EE462 MLCV 1 Lecture 3-4 Clustering (1hr) Gaussian Mixture and EM (1hr) Tae-Kyun Kim.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Gaussian Mixture Models and Expectation Maximization.
Biointelligence Laboratory, Seoul National University
Gaussian Mixture Model and the EM algorithm in Speech Recognition
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
International Conference on Intelligent and Advanced Systems 2007 Chee-Ming Ting Sh-Hussain Salleh Tian-Swee Tan A. K. Ariff. Jain-De,Lee.
Competence Centre on Information Extraction and Image Understanding for Earth Observation 29/03/07 Blind city classification using aggregation of clusterings.
CHAPTER 7: Clustering Eick: K-Means and EM (modified Alpaydin transparencies and new transparencies added) Last updated: February 25, 2014.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Max-Margin Classification of Data with Absent Features Presented by Chunping Wang Machine Learning Group, Duke University July 3, 2008 by Chechik, Heitz,
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
A DISTRIBUTION BASED VIDEO REPRESENTATION FOR HUMAN ACTION RECOGNITION Yan Song, Sheng Tang, Yan-Tao Zheng, Tat-Seng Chua, Yongdong Zhang, Shouxun Lin.
Cluster Analysis Potyó László. Cluster: a collection of data objects Similar to one another within the same cluster Similar to one another within the.
A split-and-merge framework for 2D shape summarization D. Gerogiannis, C. Nikou and A. Likas Department of Computer Science, University of Ioannina, Greece.
Gaussian Mixture Models and Expectation-Maximization Algorithm.
Prototype Classification Methods Fu Chang Institute of Information Science Academia Sinica ext. 1819
Lecture 2: Statistical learning primer for biologists
Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1.
Model-based Clustering
Design and Implementation of Speech Recognition Systems Fall 2014 Ming Li Special topic: the Expectation-Maximization algorithm and GMM Sep Some.
For multivariate data of a continuous nature, attention has focussed on the use of multivariate normal components because of their computational convenience.
Gaussian Mixture Model classification of Multi-Color Fluorescence In Situ Hybridization (M-FISH) Images Amin Fazel 2006 Department of Computer Science.
Unsupervised Learning Part 2. Topics How to determine the K in K-means? Hierarchical clustering Soft clustering with Gaussian mixture models Expectation-Maximization.
Big Data Infrastructure
Ch 12. Continuous Latent Variables ~ 12
LECTURE 11: Advanced Discriminant Analysis
Outlier Processing via L1-Principal Subspaces
Statistical Models for Automatic Speech Recognition
CS 2750: Machine Learning Expectation Maximization
Latent Variables, Mixture Models and EM
Probabilistic Models for Linear Regression
Course Outline MODEL INFORMATION COMPLETE INCOMPLETE
Bayesian Models in Machine Learning
Probabilistic Models with Latent Variables
ECE539 final project Instructor: Yu Hen Hu Fall 2005
Statistical Models for Automatic Speech Recognition
SMEM Algorithm for Mixture Models
EM for GMM.
CS 2750: Machine Learning Expectation Maximization
Unsupervised Learning II: Soft Clustering with Gaussian Mixture Models
10701 Recitation Pengtao Xie
LECTURE 21: CLUSTERING Objectives: Mixture Densities Maximum Likelihood Estimates Application to Gaussian Mixture Models k-Means Clustering Fuzzy k-Means.
LECTURE 15: REESTIMATION, EM AND MIXTURES
Biointelligence Laboratory, Seoul National University
EM Algorithm and its Applications
EM Algorithm 主講人:虞台文.
Clustering (2) & EM algorithm
Machine Learning and Data Mining Clustering
Presentation transcript:

Unsupervised-learning Methods for Image Clustering Tal Berger and Avishay Shamay Advisors: Oren Freifeld (BGU CS) and Roy Resh (Trax Image Recognition)

Project Goal Unsupervised clustering of product images. Determine the number of cluster automatically.

problem Bounding boxes are unavailable High variability in image quality and viewing angle

solution Unsupervised clustering of covariance features we implemented two algorithms for this propose: K – means GMM EM - Gaussian Mixture Model and Expectation – maximization algorithm

Data Representation ( 𝑥 1 , 𝑥 2 ,…, 𝑥 𝑛 )∈ ℝ 𝑛(𝑛+1) 2 First lets talk about what features we chose for each image and how we extract them: y {𝑥,𝑦,𝑟,𝑔,𝑏, 𝜕𝑟 𝜕𝑥 , 𝜕𝑔 𝜕𝑥 , 𝜕𝑏 𝜕𝑥 , 𝜕𝑟 𝜕𝑦 , 𝜕𝑔 𝜕𝑦 , 𝜕𝑏 𝜕𝑦 } 𝑝 11 ⋯ 𝑝 1𝑛 ⋮ ⋱ ⋮ 𝑝 𝑚1 ⋯ 𝑝_𝑚𝑛 Create covariance matrix Using covariance matrix symmetry ⋯ ⋮ ⋱ ⋮ ⋯ ( 𝑥 1 , 𝑥 2 ,…, 𝑥 𝑛 )∈ ℝ 𝑛(𝑛+1) 2 x

K - means argmin 𝑆 𝑖=1 𝑘 𝑥∈ 𝑆 𝑖 𝑥− 𝜇 𝑖 2 Given a set of observations (x1, x2, …, xn), where each observation is a d-dimensional real vector, k-means clustering aims to partition the n observations into k (≤ n) sets S = {S1, S2, …, Sk} so as to minimize the within-cluster sum of square. argmin 𝑆 𝑖=1 𝑘 𝑥∈ 𝑆 𝑖 𝑥− 𝜇 𝑖 2

GMM - EM expectation–maximization (EM) algorithm is an iterative method to find maximum likelihood estimate from incomplete data. GMM - a probabilistic model for representing the presence of subpopulations within an overall population. Model Selection to find K using BIC: 𝐵𝐼𝐶= ln 𝑛 𝐾−2 ln 𝐿 Bayesian information criterion © C. M. Bishop's book.

Conclusion and results GMM outperforms K-means. 40 < optimal K < 60 Giving higher weights to the color features improved results.

Positive results

Negative results

Q & A