Gaussian Mixture Model classification of Multi-Color Fluorescence In Situ Hybridization (M-FISH) Images Amin Fazel 2006 Department of Computer Science.

Slides:



Advertisements
Similar presentations
Unsupervised Learning
Advertisements

CS479/679 Pattern Recognition Dr. George Bebis
Face Recognition Ying Wu Electrical and Computer Engineering Northwestern University, Evanston, IL
Clustering Beyond K-means
2 – In previous chapters: – We could design an optimal classifier if we knew the prior probabilities P(wi) and the class- conditional probabilities P(x|wi)
K Means Clustering , Nearest Cluster and Gaussian Mixture
Clustering Clustering of data is a method by which large sets of data is grouped into clusters of smaller sets of similar data. The example below demonstrates.
Segmentation and Fitting Using Probabilistic Methods
Clustering II.
Clustering… in General In vector space, clusters are vectors found within  of a cluster vector, with different techniques for determining the cluster.
© University of Minnesota Data Mining for the Discovery of Ocean Climate Indices 1 CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Multiple Human Objects Tracking in Crowded Scenes Yao-Te Tsai, Huang-Chia Shih, and Chung-Lin Huang Dept. of EE, NTHU International Conference on Pattern.
Clustering.
Unsupervised Training and Clustering Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
Object Class Recognition Using Discriminative Local Features Gyuri Dorko and Cordelia Schmid.
Expectation Maximization for GMM Comp344 Tutorial Kai Zhang.
Visual Recognition Tutorial
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
EE462 MLCV 1 Lecture 3-4 Clustering (1hr) Gaussian Mixture and EM (1hr) Tae-Kyun Kim.
METU Informatics Institute Min 720 Pattern Classification with Bio-Medical Applications PART 2: Statistical Pattern Classification: Optimal Classification.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Gaussian Mixture Model and the EM algorithm in Speech Recognition
Principles of Pattern Recognition
2 Outline Introduction –Motivation and Goals –Grayscale Chromosome Images –Multi-spectral Chromosome Images Contributions Results Conclusions.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
International Conference on Intelligent and Advanced Systems 2007 Chee-Ming Ting Sh-Hussain Salleh Tian-Swee Tan A. K. Ariff. Jain-De,Lee.
CHAPTER 7: Clustering Eick: K-Means and EM (modified Alpaydin transparencies and new transparencies added) Last updated: February 25, 2014.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Learning Theory Reza Shadmehr Linear and quadratic decision boundaries Kernel estimates of density Missing data.
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
Using Feed Forward NN for EEG Signal Classification Amin Fazel April 2006 Department of Computer Science and Electrical Engineering University of Missouri.
Clustering Algorithms Presented by Michael Smaili CS 157B Spring
Quadratic Classifiers (QC) J.-S. Roger Jang ( 張智星 ) CS Dept., National Taiwan Univ Scientific Computing.
Cluster Analysis Potyó László. Cluster: a collection of data objects Similar to one another within the same cluster Similar to one another within the.
Jakob Verbeek December 11, 2009
Boosted Particle Filter: Multitarget Detection and Tracking Fayin Li.
Gaussian Mixture Models and Expectation-Maximization Algorithm.
Radial Basis Function ANN, an alternative to back propagation, uses clustering of examples in the training set.
1 Unsupervised Learning and Clustering Shyh-Kang Jeng Department of Electrical Engineering/ Graduate Institute of Communication/ Graduate Institute of.
Flat clustering approaches
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Elements of Pattern Recognition CNS/EE Lecture 5 M. Weber P. Perona.
Statistical Models for Automatic Speech Recognition Lukáš Burget.
Clustering (1) Chapter 7. Outline Introduction Clustering Strategies The Curse of Dimensionality Hierarchical k-means.
Hidden Variables, the EM Algorithm, and Mixtures of Gaussians Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 02/22/11.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Mixture Densities Maximum Likelihood Estimates.
A Study on Speaker Adaptation of Continuous Density HMM Parameters By Chin-Hui Lee, Chih-Heng Lin, and Biing-Hwang Juang Presented by: 陳亮宇 1990 ICASSP/IEEE.
Big Data Infrastructure Week 9: Data Mining (4/4) This work is licensed under a Creative Commons Attribution-Noncommercial-Share Alike 3.0 United States.
Unsupervised Learning Part 2. Topics How to determine the K in K-means? Hierarchical clustering Soft clustering with Gaussian mixture models Expectation-Maximization.
Classification of unlabeled data:
Statistical Models for Automatic Speech Recognition
Hidden Markov Models Part 2: Algorithms
REMOTE SENSING Multispectral Image Classification
DataMining, Morgan Kaufmann, p Mining Lab. 김완섭 2004년 10월 27일
Statistical Models for Automatic Speech Recognition
SMEM Algorithm for Mixture Models
Unsupervised Learning II: Soft Clustering with Gaussian Mixture Models
EE513 Audio Signals and Systems
LECTURE 21: CLUSTERING Objectives: Mixture Densities Maximum Likelihood Estimates Application to Gaussian Mixture Models k-Means Clustering Fuzzy k-Means.
Multivariate Methods Berlin Chen
Mathematical Foundations of BME
Multivariate Methods Berlin Chen, 2005 References:
EM Algorithm and its Applications
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes International.
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes International.
Presentation transcript:

Gaussian Mixture Model classification of Multi-Color Fluorescence In Situ Hybridization (M-FISH) Images Amin Fazel 2006 Department of Computer Science and Electrical Engineering University of Missouri – Kansas City

Motivation and Goals Chromosomes store genetic information Chromosome images can indicate genetic disease, cancer, radiation damage, etc. Research goals: Locate and classify each chromosome in an image Locate chromosome abnormalities Thursday, June, 2006 CS and EE Department UMKC

Karyotyping 46 human chromosomes form 24 types 22 different pairs 2 sex chromosomes, X and Y Grouped and ordered by length Banding Patterns Karyotype Thursday, June, 2006 CS and EE Department UMKC

Multi-spectral Chromosome Imaging Multiplex Fluorescence In-Situ Hybridization (M-FISH) [1996] Five color dyes (fluorophores) Each human chromosome type absorbs a unique combination of the dyes 32 (25) possible combinations of dyes distinguish 24 human chromosome types Healthy Male Thursday, June, 2006 CS and EE Department UMKC

M-FISH Images 6th dye (DAPI) binds to all chromosomes M-FISH Image 5 Dyes DAPI Channel 6th Dye Thursday, June, 2006 CS and EE Department UMKC

M-FISH Images Images of each dye obtained with appropriate optical filter Each pixel a six dimensional vector Each vector element gives contribution of a dye at pixel Chromosomal origin distinguishable at single pixel (unless overlapping) Unnecessary to estimate length, relative centromere position, or banding pattern Thursday, June, 2006 CS and EE Department UMKC

Bayesian Classification Based on probability theory A feature vector is denoted as x = [x1; x2; : : : ; xD]T D is the dimension of a vector The probability that a feature vector x belongs to class wk is p(wk|x) and this posteriori probability can be computed via and Probability density function of class wk Prior probability Thursday, June, 2006 CS and EE Department UMKC

Gaussian Probability Density Function In the D-dimensional space is the mean vector is the covariance matrix In the Gaussian distribution lies an assumption that the class model is truly a model of one basic class Thursday, June, 2006 CS and EE Department UMKC

Gaussian mixture model GMM GMM is a set of several Gaussians which try to represent groups / clusters of data therefore represent different subclasses inside one class The PDF is defined as a weighted sum of Gaussians Thursday, June, 2006 CS and EE Department UMKC

Gaussian Mixture Models Equations for GMMs: multi-dimensional case:  becomes vector ,  becomes covariance matrix . assume  is diagonal matrix: 211 1 222 233 -1 = Thursday, June, 2006 CS and EE Department UMKC

GMM Gaussian Mixture Model (GMM) is characterized by the number of components, the means and covariance matrices of the Gaussian components the weight (height) of each component Thursday, June, 2006 CS and EE Department UMKC

GMM GMM is the same dimension as the feature space (6-dimensional GMM) for visualization purposes, here are 2-dimensional GMMs: likelihood value1 value2 value2 Thursday, June, 2006 CS and EE Department UMKC

GMM These parameters are tuned using a iterative procedure called the Expectation Maximization (EM) EM algorithm: recursively updates distribution of each Gaussian model and conditional probability to increase the maximum likelihood. Thursday, June, 2006 CS and EE Department UMKC

GMM Training Flow Chart (1) Initialize the initial Gaussian means μi using the K-means clustering algorithm Initialize the covariance matrices to the distance to the nearest cluster Initialize the weights 1 / C so that all Gaussian are equally likely K-means clustering 1. Initialization: random or max. distance. 2. Search: for each training vector, find the closest code word, assign this training vector to that cell 3. Centroid Update: for each cell, compute centroid of that cell. The new code word is the centroid. 4. Repeat (2)-(3) until average distance falls below threshold Thursday, June, 2006 CS and EE Department UMKC

GMM Training Flow Chart (2) E step: Computes the conditional expectation of the complete log-likelihood, (Evaluate the posterior probabilities that relate each cluster to each data point in the conditional probability) assuming the current cluster parameters to be correct M step: Find the cluster parameters that maximize the likelihood of the data assuming that the current data distribution is correct. Thursday, June, 2006 CS and EE Department UMKC

GMM Training Flow Chart (3) recompute wn,c using the new weights, means and covariances. Stop training if wn+1,c - wn,c < threshold Or the number of epochs reach the specified value. Otherwise, continue the iterative updates. Thursday, June, 2006 CS and EE Department UMKC

GMM Test Flow Chart Present each input pattern x and compute the confidence for each class k: Where is the prior probability of class ck estimated by counting the number of training patterns Classify pattern x as the class with the highest confidence. Thursday, June, 2006 CS and EE Department UMKC

Results Training Input Data Thursday, June, 2006 CS and EE Department UMKC

Results True label Correctness Two Gaussian Correctness One Gaussian Thursday, June, 2006 CS and EE Department UMKC

Thanks for your patience ! Thursday, June, 2006 CS and EE Department UMKC