Jue Wang Michael F. Cohen IEEE CVPR 2007. Outline 1. Introduction 2. Failure Modes For Previous Approaches 3. Robust Matting 3.1 Optimized Color Sampling.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Applications of one-class classification
Bayesian Belief Propagation
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
SOFT SCISSORS: AN INTERACTIVE TOOL FOR REALTIME HIGH QUALITY MATTING International Conference on Computer Graphics and Interactive Techniques ACM SIGGRAPH.
Hongliang Li, Senior Member, IEEE, Linfeng Xu, Member, IEEE, and Guanghui Liu Face Hallucination via Similarity Constraints.
1School of CS&Eng The Hebrew University
Image Matting with the Matting Laplacian
電腦視覺 Computer and Robot Vision I Chapter2: Binary Machine Vision: Thresholding and Segmentation Instructor: Shih-Shinh Huang 1.
Graph Laplacian Regularization for Large-Scale Semidefinite Programming Kilian Weinberger et al. NIPS 2006 presented by Aggeliki Tsoli.
The General Linear Model. The Simple Linear Model Linear Regression.
Kapitel 13 “Interactive Segmentation” – p. 1 Interactive Segmentation  Live-wire approach  Random walker segmentation TexPoint fonts used in EMF. Read.
1 Roey Izkovsky Yuval Kaminka Matting Helping Superman fly since 1978.
Grayscale Image Matting And Colorization Tongbo Chen, Yan Wang, Volker Schillings, Christoph Meinel FB IV-Informatik, University of Trier, Trier 54296,
Texture Segmentation Based on Voting of Blocks, Bayesian Flooding and Region Merging C. Panagiotakis (1), I. Grinias (2) and G. Tziritas (3)
Image Matting and Its Applications Chen-Yu Tseng Advisor: Sheng-Jyh Wang
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
Content Based Image Clustering and Image Retrieval Using Multiple Instance Learning Using Multiple Instance Learning Xin Chen Advisor: Chengcui Zhang Department.
Boundary matting for view synthesis Samuel W. Hasinoff Sing Bing Kang Richard Szeliski Computer Vision and Image Understanding 103 (2006) 22–32.
Prénom Nom Document Analysis: Parameter Estimation for Pattern Recognition Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Background Removal of Multiview Images by Learning Shape Priors Yu-Pao Tsai, Cheng-Hung Ko, Yi-Ping Hung, and Zen-Chung Shih.
CS 376b Introduction to Computer Vision 04 / 08 / 2008 Instructor: Michael Eckmann.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Cluster Analysis.  What is Cluster Analysis?  Types of Data in Cluster Analysis  A Categorization of Major Clustering Methods  Partitioning Methods.
Adaptive Rao-Blackwellized Particle Filter and It’s Evaluation for Tracking in Surveillance Xinyu Xu and Baoxin Li, Senior Member, IEEE.
A Closed Form Solution to Natural Image Matting
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Ensemble Learning: An Introduction
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Optical flow and Tracking CISC 649/849 Spring 2009 University of Delaware.
Abstract Extracting a matte by previous approaches require the input image to be pre-segmented into three regions (trimap). This pre-segmentation based.
Introduction to Boosting Aristotelis Tsirigos SCLT seminar - NYU Computer Science.
An Iterative Optimization Approach for Unified Image Segmentation and Matting Hello everyone, my name is Jue Wang, I’m glad to be here to present our paper.
Linear Discriminant Functions Chapter 5 (Duda et al.)
Graph-based consensus clustering for class discovery from gene expression data Zhiwen Yum, Hau-San Wong and Hongqiang Wang Bioinformatics, 2007.
CHAMELEON : A Hierarchical Clustering Algorithm Using Dynamic Modeling
Geo479/579: Geostatistics Ch13. Block Kriging. Block Estimate  Requirements An estimate of the average value of a variable within a prescribed local.
PATTERN RECOGNITION AND MACHINE LEARNING
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
06 - Boundary Models Overview Edge Tracking Active Contours Conclusion.
Efficient Gathering of Correlated Data in Sensor Networks
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
1/20 Obtaining Shape from Scanning Electron Microscope Using Hopfield Neural Network Yuji Iwahori 1, Haruki Kawanaka 1, Shinji Fukui 2 and Kenji Funahashi.
Mean-shift and its application for object tracking
7.1. Mean Shift Segmentation Idea of mean shift:
EE 492 ENGINEERING PROJECT LIP TRACKING Yusuf Ziya Işık & Ashat Turlibayev Yusuf Ziya Işık & Ashat Turlibayev Advisor: Prof. Dr. Bülent Sankur Advisor:
Finite Element Method.
Object Stereo- Joint Stereo Matching and Object Segmentation Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on Michael Bleyer Vienna.
Texture. Texture is an innate property of all surfaces (clouds, trees, bricks, hair etc…). It refers to visual patterns of homogeneity and does not result.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Data Extraction using Image Similarity CIS 601 Image Processing Ajay Kumar Yadav.
Presenter : Kuang-Jui Hsu Date : 2011/3/24(Thur.).
Segmentation of Vehicles in Traffic Video Tun-Yu Chiang Wilson Lau.
Gaussian Mixture Models and Expectation-Maximization Algorithm.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Review of fundamental 1 Data mining in 1D: curve fitting by LLS Approximation-generalization tradeoff First homework assignment.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Chapter1: Introduction Chapter2: Overview of Supervised Learning
Improving Image Matting using Comprehensive Sampling Sets CVPR2013 Oral.
Thresholding Foundation:. Thresholding In A: light objects in dark background To extract the objects: –Select a T that separates the objects from the.
Implementing the By: Matthew Marsh Supervisors: Prof Shaun Bangay Mrs Adele Lobb segmentation technique as a plugin for the GIMP.
A Binary Linear Programming Formulation of the Graph Edit Distance Presented by Shihao Ji Duke University Machine Learning Group July 17, 2006 Authors:
Dimension reduction (2) EDR space Sliced inverse regression Multi-dimensional LDA Partial Least Squares Network Component analysis.
LINEAR CLASSIFIERS The Problem: Consider a two class task with ω1, ω2.
Iterative Optimization
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
“grabcut”- Interactive Foreground Extraction using Iterated Graph Cuts
Presented by: Chang Jia As for: Pattern Recognition
“Traditional” image segmentation
Presentation transcript:

Jue Wang Michael F. Cohen IEEE CVPR 2007

Outline 1. Introduction 2. Failure Modes For Previous Approaches 3. Robust Matting 3.1 Optimized Color Sampling 3.2 Collecting the Sample Set 3.3 Matte Optimization Solving for Optimal αs 4. Result5. Conclusion

1. Introduction  Matting refers to the problem of soft and accurate foreground extraction from an image. C is formulated as a convex combination of a foreground image F and a background image B as C z = α z F z + (1 − α z )B z …………………(1) *C is input image, F and B are respectively foreground and background image * z refers to pixel locations, and α z is the foreground opacity of the pixel.

1. Introduction  Roughly speaking, previous matting approaches can be classified into two categories based on how they make use of natural image statistics. 1. sampling-based approaches─ assume that the foreground and background colors of an unknown pixel can be explicitly estimated by examining nearby pixels that have been specified by the user as foreground or background. (ex: KnockOut2 matting, Bayesian matting, Belief Propagation matting…etc) 2. propagation-based approaches ─ assume foreground and background colors are locally smooth, for example,that they can be modelled as constant or linearly varying. (ex: Poisson matting, Closed-form matting…etc)

2. Failure Modes For Previous Approaches  For sampling-based approaches pixel P A fits the linear model very well, thus it has a high probability of being a true mixed pixel between the foreground and background clusters. On the contrary, pixel P B is far away from the interpolation line thus it is very unlikely to be generated by a linear combination of the two clusters. # previous approaches ignore this fact and simply estimate an alpha value for P B based on its projection to the line P’ B

2. Failure Modes For Previous Approaches  For propagation-based approaches For complex foreground and background patterns, samples collected from local regions do not have a uniform color distribution, as shown in above Figure. In this case, propagation-based approaches will fail (at least partially) since the smoothness assumption is violated.

2. Failure Modes For Previous Approaches Figure 2a shows the original image with the user specified foreground (red) and unknown region (yellow). Although the foreground color is relatively uniform, the background contains complex patterns, in which dark regions match well with the foreground color. Figure 2 b, c, and d show that other matting will produce noticeable artifacts due to color sampling errors or assumption violations.

3. Robust Matting  This algorithm is based on an optimized color sampling scheme, and Random Walk optimizer is employed to solve for the matte.

3.1 Optimized Color Sampling Precondition: Given the input image and a roughly specified trimap. ( trimap can separate from foreground 、 background and unknow region)  For a pixel z with unknown α assembles a large number of foreground and background samples as candidates for estimating the true foreground and background colors at this location.

3.1 Optimized Color Sampling  pick out “good” samples from this large candidate set. =>Good sample pairs should explain any mixed foreground/background pixels as linear combinations of the samples. Assume that B j is a simple in B, F i is a simple in F. =>B j and F i in this Figure are good samples for P A, but not for P B.

3.1 Optimized Color Sampling for a pair of foreground and background colors F i and B j, the estimated alpha value is …………..(2)

3.1 Optimized Color Sampling And We define a distance ratio R d (F i,B j ), which evaluates this sample pair by examining the ratio of the distances between (1) the pixel color, C, and the color it would have,, predicted by the linear model in Equation 1, and (2) the distance between the foreground/background pair. ……….(3) In the example shown in Figure, the distance ratio will be much higher for P B than P A, indicating the samples are not as good for estimating alpha for P B.

3.1 Optimized Color Sampling The distance ratio alone will favor sample pairs that are widely spread in color space since the denominator ||F i −B j || will be large.

3.1 Optimized Color Sampling We expect most pixels to be fully foreground or background, pixels with colors that lie nearby in color space to foreground and background samples are more likely to be fully foreground or background themselves. Thus, for each individual sample we define two more weights w(F i ) and w(B j ) as W(F i ) = exp { ─ || F i – C || 2 / D 2 F }……………(4) and W(B j ) = exp { ─ || B j – C || 2 / D 2 B }…………...(5) ** where D F and D B are the minimum distances between foreground/background sample and the current pixel,(i.e, min i ( || F i – C || ) and min j ( || B j – C || ).

3.1 Optimized Color Sampling Combining these factors( i.e R d (F i,B j ) 、 W(F i ) 、 W(B j )), we calculate a final confidence value f(F i,B j ) for a sample pair as …..(6) **where σ=0.1 in this system. We examine the confidence of every pair of foreground and background samples. Finally, we select a small number of pairs (3 in our system) with the highest confidences. The average estimated alpha value and confidence of these three sample pairs are taken as the final values in the color sampling step.

3.2 Collecting the Sample Set  Previous approaches such as Bayesian matting and Belief Propagation matting collect pixels known to be fully foreground or background that have the shortest spatial distances to the target pixel as samples, but may not fully span the variation in foreground and background colors.  In Robust Sampling Method spread the sampling of foreground and background samples along the boundaries of known foreground and background regions. In this way the sample set can better capture the variation of foreground and background colors.

3.3 Matte Optimization  sampling process leads to a good initial alpha estimate and a confidence value for each pixel,and this initial estimate can be further improved by leveraging a priori expectations about more global aspects of the alpha matte. In particular, we expect the matte to exhibit local smoothness. We also expect alpha values of one or zero (fully foreground or background) to be much more common than mixed pixels.

3.3 Matte Optimization  Our expectation for the matte is thus two fold: firstly, it should respect the alphas chosen for each individual pixel (data constraint) especially when the confidence value is high; secondly, the matte should be locally smooth (neighborhood constraint). →This expectation can be satisfied by solving a graph labelling problem in previous graph-based image labelling approaches.

3.3 Matte Optimization  Graph labelling Problem  Ω F 、 Ω B → virtual nodes representing pure foreground and pure background.  white nodes → unknown pixels on the image lattice.  light red and light blue nodes → known pixels marked by the user.  data weight → defined between each pixel and a virtual node to enforce the data constraint. ( ex: W i,F 、 W i,B )  edge weight → defined between two neighboring pixels to enforce the neighborhood constraint. ( ex: W i,j )

3.3 Matte Optimization  Data weights correspond to relative probabilities of a node being foreground or background.  For nodes with high confidence values,, we rely on the alpha that fits the linear model from the selected samples.  For nodes with low confidence, we have a higher expectation that the node is fully foreground or background. By above rules,for an unknown pixel i, two data weights W( i, F) and W( i,B),They are defined as and ……………(7) ** and are estimated alpha and confidence values,and is boolean function returning 0 or 1, is a free parameter in our system which balances the data weight and the edge weight.

3.3 Matte Optimization  Edge weights between nodes to encourage alpha to have local smoothness. They reference the closed-form matting system [A.Levin,et.al 06] sets the weights between neighboring pixels based on their color difference computed from local color distributions. the neighborhood term, W ij, is defined by a sum over all 3 × 3 windows that contain pixels i and j. …….(8) ** represents the set of 3 × 3 windows containing pixels i and j, and k iterates over those windows. and are the color mean and variance in each window. is a regularization coefficient which is set to be 10 −5 in our system.

3.3.1 Solving for Optimal α s  The fact that alpha values are continuous, we avoid discrete labelling optimizations such as graph cut. So,we solve the graph labelling problem as a Random Walk, which has been shown to minimize the total graph energy over real values. First, we construct a Laplacian matrix for the graph as where W ii =. L is thus a sparse, symmetric, positive- definite matrix with dimension N ×N, where N is the number of all nodes in the graph, including all pixels in the image plus two virtual nodes Ω B and Ω F.

3.3.1 Solving for Optimal α s We then decompose L into blocks corresponding to unknown nodes P u, and known nodes P k including user labelled pixels and virtual nodes, as It has been shown [L. Grady. 06] that the probabilities of unknown pixels belonging to a certain label (for example, foreground) is the solution to L u A u = - R T A K …………..(11) A u is the vector of unknown alphas we wish to solve for, and A k is the vector encoding the boundary conditions, (i.e., 1’s and 0’s for the known alpha values of the virtual and user specified nodes). Lu is guaranteed to be nonsingular for a connected graph, thus the solution Au is guaranteed to exist and be unique with values guaranteed to lie between 0 and 1. We use Conjugate Gradient (CG) to solve the linear system.

4. Result

** I a as the minimal MSE value, and I r as the difference between the maximal and minimal values.

4. Result

5. Conclusion  Robust matting approach that combines the advantages of sampling-based approaches and propagation-based approaches. We propose an optimized color sampling method, which explicitly avoids the weak assumptions of previous approaches, thus enable our algorithm to generate accurate mattes in a robust way for complex images.