Robust Principal Components Analysis IT530 Lecture Notes.

Slides:



Advertisements
Similar presentations
5.4 Basis And Dimension.
Advertisements

General Linear Model With correlated error terms  =  2 V ≠  2 I.
Applied Informatics Štefan BEREŽNÝ
Compressive Sensing IT530, Lecture Notes.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
T HE POWER OF C ONVEX R ELAXATION : N EAR - OPTIMAL MATRIX COMPLETION E MMANUEL J. C ANDES AND T ERENCE T AO M ARCH, 2009 Presenter: Shujie Hou February,
CS345 Data Mining Link Analysis Algorithms Page Rank Anand Rajaraman, Jeffrey D. Ullman.
Lighting affects appearance. What is the Question ? (based on work of Basri and Jacobs, ICCV 2001) Given an object described by its normal at each.
Foreground Modeling The Shape of Things that Came Nathan Jacobs Advisor: Robert Pless Computer Science Washington University in St. Louis.
Jue Wang Michael F. Cohen IEEE CVPR Outline 1. Introduction 2. Failure Modes For Previous Approaches 3. Robust Matting 3.1 Optimized Color Sampling.
Bayesian Robust Principal Component Analysis Presenter: Raghu Ranganathan ECE / CMR Tennessee Technological University January 21, 2011 Reading Group (Xinghao.
Visual Recognition Tutorial
Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, Microsoft Research Asia 3 University of California.
Shadow removal algorithms Shadow removal seminar Pavel Knur.
1 In-Network PCA and Anomaly Detection Ling Huang* XuanLong Nguyen* Minos Garofalakis § Michael Jordan* Anthony Joseph* Nina Taft § *UC Berkeley § Intel.
Maximum likelihood Conditional distribution and likelihood Maximum likelihood estimations Information in the data and likelihood Observed and Fisher’s.
2.III. Basis and Dimension 1.Basis 2.Dimension 3.Vector Spaces and Linear Systems 4.Combining Subspaces.
Math for CSLecture 41 Linear Least Squares Problem Over-determined systems Minimization problem: Least squares norm Normal Equations Singular Value Decomposition.
Minimaxity & Admissibility Presenting: Slava Chernoi Lehman and Casella, chapter 5 sections 1-2,7.
Maximum likelihood (ML) and likelihood ratio (LR) test
Random Convolution in Compressive Sampling Michael Fleyer.
Preference Analysis Joachim Giesen and Eva Schuberth May 24, 2006.
Camera Calibration CS485/685 Computer Vision Prof. Bebis.
化工應用數學 授課教師: 郭修伯 Lecture 9 Matrices
AMSC 6631 Sparse Solutions of Linear Systems of Equations and Sparse Modeling of Signals and Images: Midyear Report Alfredo Nava-Tudela John J. Benedetto,
Linear Algebra Lecture 25.
Matrix Completion IT530 Lecture Notes.
Cs: compressed sensing
A matrix equation has the same solution set as the vector equation which has the same solution set as the linear system whose augmented matrix is Therefore:
Recovering low rank and sparse matrices from compressive measurements Aswin C Sankaranarayanan Rice University Richard G. Baraniuk Andrew E. Waters.
Sparse Matrix Factorizations for Hyperspectral Unmixing John Wright Visual Computing Group Microsoft Research Asia Sept. 30, 2010 TexPoint fonts used in.
AN ORTHOGONAL PROJECTION
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Orthogonalization via Deflation By Achiya Dax Hydrological Service Jerusalem, Israel
Efficient computation of Robust Low-Rank Matrix Approximations in the Presence of Missing Data using the L 1 Norm Anders Eriksson and Anton van den Hengel.
Section 2.3 Properties of Solution Sets
Introduction to Video Background Subtraction 1. Motivation In video action analysis, there are many popular applications like surveillance for security,
4.8 Rank Rank enables one to relate matrices to vectors, and vice versa. Definition Let A be an m  n matrix. The rows of A may be viewed as row vectors.
E XACT MATRIX C OMPLETION VIA CONVEX OPTIMIZATION E MMANUEL J. C ANDES AND B ENJAMIN R ECHT M AY 2008 Presenter: Shujie Hou January, 28 th,2011 Department.
A Flexible New Technique for Camera Calibration Zhengyou Zhang Sung Huh CSPS 643 Individual Presentation 1 February 25,
Elementary Linear Algebra Anton & Rorres, 9 th Edition Lecture Set – 07 Chapter 7: Eigenvalues, Eigenvectors.
Class 24: Question 1 Which of the following set of vectors is not an orthogonal set?
Review of fundamental 1 Data mining in 1D: curve fitting by LLS Approximation-generalization tradeoff First homework assignment.
1 a1a1 A1A1 a2a2 a3a3 A2A Mixed Strategies When there is no saddle point: We’ll think of playing the game repeatedly. We continue to assume that.
Large-Scale Matrix Factorization with Missing Data under Additional Constraints Kaushik Mitra University of Maryland, College Park, MD Sameer Sheoreyy.
3D Reconstruction Using Image Sequence
Lighting affects appearance. Lightness Digression from boundary detection Vision is about recovery of properties of scenes: lightness is about recovering.
Camera Model Calibration
Link Analysis Algorithms Page Rank Slides from Stanford CS345, slightly modified.
Dynamic Background Learning through Deep Auto-encoder Networks Pei Xu 1, Mao Ye 1, Xue Li 2, Qihe Liu 1, Yi Yang 2 and Jian Ding 3 1.University of Electronic.
Finite Element Method. History Application Consider the two point boundary value problem.
Linear Programming Chap 2. The Geometry of LP  In the text, polyhedron is defined as P = { x  R n : Ax  b }. So some of our earlier results should.
Chapter 5 Chapter Content 1. Real Vector Spaces 2. Subspaces 3. Linear Independence 4. Basis and Dimension 5. Row Space, Column Space, and Nullspace 6.
MathematicalMarketing Slide 5.1 OLS Chapter 5: Ordinary Least Square Regression We will be discussing  The Linear Regression Model  Estimation of the.
Graphics Lecture 14: Slide 1 Interactive Computer Graphics Lecture 14: Radiosity - Computational Issues.
Lecture XXVII. Orthonormal Bases and Projections Suppose that a set of vectors {x 1,…,x r } for a basis for some space S in R m space such that r  m.
SketchVisor: Robust Network Measurement for Software Packet Processing
REVIEW Linear Combinations Given vectors and given scalars
Orthogonal Subspace Projection - Matched Filter
Structure from motion Input: Output: (Tomasi and Kanade)
Singular Value Decomposition
Sparse and Redundant Representations and Their Applications in
Basis and Dimension Basis Dimension Vector Spaces and Linear Systems
Sparse and Redundant Representations and Their Applications in
Linear Algebra Lecture 39.
2.III. Basis and Dimension
Rank-Sparsity Incoherence for Matrix Decomposition
I.4 Polyhedral Theory (NW)
I.4 Polyhedral Theory.
Structure from motion Input: Output: (Tomasi and Kanade)
Presentation transcript:

Robust Principal Components Analysis IT530 Lecture Notes

Basic Question Consider a matrix M of size n1 x n2 that is the sum of two components – L (a low-rank components) and S (a component with sparse but unknown support). Can we recover L and S given only M? Note: support of S means all those matrix entries (i,j) where S(i,j) ≠ 0.

Why ask this question? Scenario 1 Consider a video taken with a static camera (like at airports or on overhead bridges on expressways). Such videos contain a “background” portion that is static or changes infrequently, and a moving portion called “foreground” (people, cars etc.). In many applications (esp. surveillance for security or traffic monitoring), we need to detect and track the moving foreground.

Why ask this question? Scenario 1 Let us represent this video as a N x T matrix, where the number of columns T equals the number of frames in the video, and where each column containing a single video-frame of N = N1 N2 pixels converted to a vector form. This matrix can clearly be expressed as the sum of a low-rank background matrix and a sparse foreground matrix.

Why ask this question? Scenario 2 It is well-known that the images of a convex Lambertian object taken from the same viewpoint, but under K different lighting directions, span a low-dimensional space (ref: Basri and Jacobs, “Lambertian reflectance and linear subspaces”, IEEE TPAMI, 2003). This means that a matrix M of size N x K (each column of which is an image captured under one of the K lighting directions and represented in vectorized form) has a low approximate rank – in fact at most rank 9.

Why ask this question? Scenario 2 But many objects of interest (such as human faces) are neither convex nor exactly Lambertian. Effects such as shadows, specularities or saturation cause violation of the low-rank assumption. But these errors have sparse spatial support, even though their magnitudes are large. Here again, M can be represented as a low-rank background and a sparse foreground.

Theorem 1 (Informal Statement) Consider a matrix M of size n 1 by n 2 which is the sum of a “sufficiently low-rank” component L and a “sufficiently sparse” component S whose support is uniformly randomly distributed in the entries of M. Then the solution of the following optimization problem (known as principal component pursuit) yields exact estimates of L and S with “very high” probability:

Be careful! There is an implicit assumption in here that the low-rank component is not sparse (e.g.: a matrix containing zeros everywhere except the right top corner won’t do!). And that the sparse component is not low- rank (that’s why we said that it’s support is drawn from a uniform random distribution).

Be careful! In addition, the low-rank component L must obey the assumptions A0 and A1 (required in matrix completion theory).

Theorem 1 (Formal Statement)

Comments on Theorem 1 If the value m is smaller, the upper bound on the maximum rank allowed is larger. We require the support of S to be sparse and uniformly distributed but make no assumption about the magnitude/sign of the non-zero elements in S. But we don’t know the support of S. In the matrix completion problem, we knew which entries in the matrix were “missing”. Here we do not know the support of S. And the entries in S are not “missing” but unknown and grossly corrupted!

Matrix Recovery with Corrupted as well as missing data In some cases, we may have missing entries in the matrix (with known location) in addition to corrupted entries (whose location and magnitude we DON’T know). In such cases, we solve:

Theorem 2 (Formal Statement)

Comments on Theorem 2 If all entries were observed (i.e. m = n1n2), then this reduces to Theorem 1. If t = 0, this becomes a pure matrix completion problem (no sparse component), but which is solved using a slightly different objective function (question: what’s that difference?). In the t = 0 case, the minimization routine is guaranteed to give us a low-rank component as usual, but a “sparse” component that contains all zeros.

Results