Data-Driven Markov Chain Monte Carlo Presented by Tomasz MalisiewiczTomasz Malisiewicz for Advanced PerceptionAdvanced Perception 3/1/2006.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

CSCE643: Computer Vision Bayesian Tracking & Particle Filtering Jinxiang Chai Some slides from Stephen Roth.
Ch 11. Sampling Models Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by I.-H. Lee Biointelligence Laboratory, Seoul National.
Computer Vision Lab. SNU Young Ki Baik An Introduction to MCMC for Machine Learning (Markov Chain Monte Carlo)
Ghunhui Gu, Joseph J. Lim, Pablo Arbeláez, Jitendra Malik University of California at Berkeley Berkeley, CA
São Paulo Advanced School of Computing (SP-ASC’10). São Paulo, Brazil, July 12-17, 2010 Looking at People Using Partial Least Squares William Robson Schwartz.
Introduction to Sampling based inference and MCMC Ata Kaban School of Computer Science The University of Birmingham.
CHAPTER 16 MARKOV CHAIN MONTE CARLO
Image Parsing: Unifying Segmentation and Detection Z. Tu, X. Chen, A.L. Yuille and S-C. Hz ICCV 2003 (Marr Prize) & IJCV 2005 Sanketh Shetty.
Bayesian statistics – MCMC techniques
BAYESIAN INFERENCE Sampling techniques
Visual Recognition Tutorial
EE-148 Expectation Maximization Markus Weber 5/11/99.
1 Hierarchical Image-Motion Segmentation using Swendsen-Wang Cuts Adrian Barbu Siemens Corporate Research Princeton, NJ Acknowledgements: S.C. Zhu, Y.N.
Mean Shift A Robust Approach to Feature Space Analysis Kalyan Sunkavalli 04/29/2008 ES251R.
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
Segmentation and Tracking of Multiple Humans in Crowded Environments Tao Zhao, Ram Nevatia, Bo Wu IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE,
1 On the Statistical Analysis of Dirty Pictures Julian Besag.
Segmentation Divide the image into segments. Each segment:
Adaptive Rao-Blackwellized Particle Filter and It’s Evaluation for Tracking in Surveillance Xinyu Xu and Baoxin Li, Senior Member, IEEE.
1 Integration of Background Modeling and Object Tracking Yu-Ting Chen, Chu-Song Chen, Yi-Ping Hung IEEE ICME, 2006.
Today Introduction to MCMC Particle filters and MCMC
Visual Recognition Tutorial
Amos Storkey, School of Informatics. Density Traversal Clustering and Generative Kernels a generative framework for spectral clustering Amos Storkey, Tom.
Learning Bayesian Networks
Computer vision: models, learning and inference Chapter 10 Graphical Models.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Monte Carlo Methods in Partial Differential Equations.
Tal Mor  Create an automatic system that given an image of a room and a color, will color the room walls  Maintaining the original texture.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Computer Vision James Hays, Brown
Overview G. Jogesh Babu. Probability theory Probability is all about flip of a coin Conditional probability & Bayes theorem (Bayesian analysis) Expectation,
 1  Outline  stages and topics in simulation  generation of random variates.
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
Mean-shift and its application for object tracking
Mean Shift Theory and Applications Reporter: Zhongping Ji.
Stochastic Algorithms Some of the fastest known algorithms for certain tasks rely on chance Stochastic/Randomized Algorithms Two common variations – Monte.
A General Framework for Tracking Multiple People from a Moving Camera
CSE 185 Introduction to Computer Vision Pattern Recognition 2.
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
Supervised Learning of Edges and Object Boundaries Piotr Dollár Zhuowen Tu Serge Belongie.
EECS 274 Computer Vision Segmentation by Clustering II.
Randomized Algorithms for Bayesian Hierarchical Clustering
CS654: Digital Image Analysis Lecture 30: Clustering based Segmentation Slides are adapted from:
Probabilistic Latent Query Analysis for Combining Multiple Retrieval Sources Rong Yan Alexander G. Hauptmann School of Computer Science Carnegie Mellon.
Boosted Particle Filter: Multitarget Detection and Tracking Fayin Li.
1 Tree Crown Extraction Using Marked Point Processes Guillaume Perrin Xavier Descombes – Josiane Zerubia ARIANA, joint research group CNRS/INRIA/UNSA INRIA.
 Present by 陳群元.  Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental.
Visual Tracking by Cluster Analysis Arthur Pece Department of Computer Science University of Copenhagen
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Segmentation Through Optimization Pyry Matikainen.
Introduction to Sampling Methods Qi Zhao Oct.27,2004.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Hidden Variables, the EM Algorithm, and Mixtures of Gaussians Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 02/22/11.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 17: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
Bayesian Hierarchical Clustering Paper by K. Heller and Z. Ghahramani ICML 2005 Presented by David Williams Paper Discussion Group ( )
Bayesian Inference and Visual Processing: Image Parsing & DDMCMC. Alan Yuille (Dept. Statistics. UCLA) Tu, Chen, Yuille & Zhu (ICCV 2003).
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Introduction to Sampling based inference and MCMC
Advanced Statistical Computing Fall 2016
CSC321: Neural Networks Lecture 19: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
Jun Liu Department of Statistics Stanford University
Data-Driven Markov Chain Monte Carlo
Outline Image Segmentation by Data-Driven Markov Chain Monte Carlo
Edges/curves /blobs Grammars are important because:
Image Parsing & DDMCMC. Alan Yuille (Dept. Statistics. UCLA)
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
EM Algorithm and its Applications
Presentation transcript:

Data-Driven Markov Chain Monte Carlo Presented by Tomasz MalisiewiczTomasz Malisiewicz for Advanced PerceptionAdvanced Perception 3/1/2006

Overview of Talk What is Image Segmentation? How to find a good segmentation? DDMCMC results Image segmentation in a Bayesian statistical framework Markov Chain Monte Carlo for exploring the space of all segmentations Data-Driven methods for exploiting image data and speeding up MCMC

DDMCMC Motivation Iterative approach: consider many different segmentations and keep the good ones Few tunable parameters, ex) # of segments encoded into prior DDMCMC vs Ncuts

Berkeley Segmentation Database Image Berkeley Ncuts K=30DDMCMC

Why a rigorous formulation? Allows us to define what we want the segmentation algorithm to return Assigning a Score to a segmentation

Formulation #1 (and you thought you knew what image segmentation was) Image Lattice: Image: For any point either or Lattice partition into K disjoint regions: Region is discrete label map: Region Boundary is Continuous: An image partition into disjoint regions is not An image segmentation! Regions Contents Are Key!

Formulation #2 (and you thought you knew what image segmentation was) Each Image Region is a realization from a probabilistic model are parameters of model indexed by A segmentation is denoted by a vector of hidden variables W; K is number of regions Bayesian Framework: Space of all segmentations Prior LikelihoodPosterior

Prior over segmentations (do you like exponentials?) ~ uniform # of model params Want less regions Want round-ish regions Want small regions Want less complex models

Likelihood for Images Visual Patterns are independent stochastic processes is model-type index is model parameter vector is image appearance in i-th region Grayscale Color

Four Gray-level Models Uniform Clutter Texture Shading Gray-level model space: GaussianIntensity Histogram FB Response Histogram B-Spline

Three Color Models (L*,u*,v*) Gaussian Mixture of 2 Gaussians Bezier Spline Color model space:

Calibration Likelihoods are calibrated using empirical study Calibration required to make likelihoods for different models comparable (necessary for model competition) Principled? or Hack?

What did we just do? Def. of Segmentation: Score (probability) of Segmentation: Likelihood of Image = product of region likelihoods Regions defined by k-partition:

What do we do with scores? Search

Search through what? Anatomy of Solution Space Space of all k-partitions General partition space Space of all segmentations Partition space K Model spaces Scene Space or

Searching through segmentations Exhaustive Enumeration of all segmentations Greedy Search (Gradient Ascent) Stochastic Search MCMC based exploration Takes too long! Local minima! Takes too long Described in the rest of this talk!

Why MCMC What is it? What does it do? -A clever way of searching through a high-dimensional space -A general purpose technique of generating samples from a probability -Iteratively searches through space of all segmentations by constructing a Markov Chain which converges to stationary distribution

Designing Markov Chains Three Markov Chain requirements Ergodic: from an initial segmentation W 0, any other state W can be visited in finite time (no greedy algorithms); ensured by jump-diffusion dynamics Aperiodic: ensured by random dynamics Detailed Balance: every move is reversible

5 Dynamics 1.) Boundary Diffusion 2.) Model Adaptation 3.) Split Region 4.) Merge Region 5.) Switch Region Model At each iteration, we choose a dynamic with probability q(1),q(2),q(3),q(4),q(5)

Dynamics 1: Boundary Diffusion Diffusion* within Boundary Between Regions i and j Brownian Motion Along Curve Normal Temperature Decreases over Time *Movement within partition space

Dynamics 2: Model Adaptation Fit the parameters* of a region by steepest ascent *Movement within cue space

Dynamics 3-4: Split and Merge Split one region into two Remaining Variables Are unchanged Probability of Proposed Split Conditional Probability of how likely chain proposes to move to W’ from W Data-Driven Speedup

Dynamics 3-4: Split and Merge Merge two Regions Remaining Variables Are unchanged Probability of Proposed Merge Data-Driven Speedup

Dynamics 5: Model Switching Change models Proposal Probabilities Data-Driven Speedup

Motivation of DD Region Splitting: How to decide where to split a region? Model Switching: Once we switch to a new model, what parameters do we jump to? vs Model Adaptation Required some initial parameter vector

Data Driven Methods Focus on boundaries and model parameters derived from data: compute these before MCMC starts Cue Particles: Clustering in Model Space K-partition Particles: Edge Detection Particles Encode Probabilities Parzen Window Style

Cue Particles In Action Clustering in Color Space

Cue Particles Extract Feature at each point in image m weighted cue particles are the output of a clustering algorithm Model Index Saliency Map Probability that Feature belongs To cluster

K-partition Particles in Action Edge detection gives us a good idea of where we expect a boundary to be located

K-partition Particles Edge detection and tracing at 3 scales Partition Map consists of “metaregions” Metaregions are used to construct regions is the set of all k-partitions based on

K-partition Particles is the set of all k-partitions based on Each in is a k-partition particle in partition space

Particles or Parzen Window* Locations? What is this particle business about? A particle is just the position of a parzen- window which is used for density estimation 1D particles *Parzen Windowing also known as: Kernel Density Estimation, Non-parametric density estimation

Nonparametric Probability Densities in Cue Spaces Weighted cue particles encode nonparametric probability density in G(x) is a parzen-window centered at 0 is computed once for each image is computed at run-time

Nonparametric Probability Densities in Partition Spaces Each k-partition particle has uniform weight and encodes nonparametric probability density in partition space Using all scales

Are you awake: What did we just do? Scores (Probability of Segmentation)  Search 5 MCMC dynamics Data-Driven Speedup (key to making MCMC work in finite time) So what type of answer does the Markov Chain return? What can we do with this answer? How many answers to we want?

Multiple Solutions MAP gives us one solution Output of MCMC sampling How do we get multiple solutions? Parzen Windows: Again Scene Particles

Why multiple solutions? Segmentation is often not the final stage of computation A higher level task such as recognition can utilize a segmentation We don’t want to make any hard decision before recognition multiple segmentations = good idea

K-adventurers We want to keep a fixed number K of segmentations but we don’t want to keep trivially different segmentations Goal: Keep the K segmentations that best preserve the posterior probability in KL-sense Greedy Algorithm: - Add new particle, remove worst particle

Results (Multiple Solutions)

Results

Results (Color Images)

Conclusions DDMCMC: Combines Generative (top- down) and Discriminative (bottom-up) approaches Traverse the space of all segmentations via Markov Chains Does your head hurt? Questions?

References DDMCMC Paper: pers/tu-pami-02.pdf pers/tu-pami-02.pdf DDMCMC Website: MCMC_segmentation.htm MCMC_segmentation.htm MCMC Tutorial by Authors: m m