Nuclear Norm Heuristic for Rank Minimization

Slides:



Advertisements
Similar presentations
Sublinear-time Algorithms for Machine Learning Ken Clarkson Elad Hazan David Woodruff IBM Almaden Technion IBM Almaden.
Advertisements

Duality for linear programming. Illustration of the notion Consider an enterprise producing r items: f k = demand for the item k =1,…, r using s components:
Globally Optimal Estimates for Geometric Reconstruction Problems Tom Gilat, Adi Lakritz Advanced Topics in Computer Vision Seminar Faculty of Mathematics.
T HE POWER OF C ONVEX R ELAXATION : N EAR - OPTIMAL MATRIX COMPLETION E MMANUEL J. C ANDES AND T ERENCE T AO M ARCH, 2009 Presenter: Shujie Hou February,
by Rianto Adhy Sasongko Supervisor: Dr.J.C.Allwright
Exact or stable image\signal reconstruction from incomplete information Project guide: Dr. Pradeep Sen UNM (Abq) Submitted by: Nitesh Agarwal IIT Roorkee.
Network Systems Lab. Korea Advanced Institute of Science and Technology No.1 Some useful Contraction Mappings  Results for a particular choice of norms.
Separating Hyperplanes
Inexact SQP Methods for Equality Constrained Optimization Frank Edward Curtis Department of IE/MS, Northwestern University with Richard Byrd and Jorge.
Visual Recognition Tutorial
Numerical Optimization
UMass Lowell Computer Science Analysis of Algorithms Prof. Karen Daniels Fall, 2006 Lecture 9 Wednesday, 11/15/06 Linear Programming.
Computational Geometry The art of finding algorithms for solving geometrical problems Literature: –M. De Berg et al: Computational Geometry, Springer,
Dual Problem of Linear Program subject to Primal LP Dual LP subject to ※ All duality theorems hold and work perfectly!
Support Vector Classification (Linearly Separable Case, Primal) The hyperplanethat solves the minimization problem: realizes the maximal margin hyperplane.
Martin Burger Institut für Numerische und Angewandte Mathematik European Institute for Molecular Imaging CeNoS Total Variation and related Methods Numerical.
Constrained Optimization
3D reconstruction of cameras and structure x i = PX i x’ i = P’X i.
Classification Problem 2-Category Linearly Separable Case A- A+ Malignant Benign.
1 Computational Learning Theory and Kernel Methods Tianyi Jiang March 8, 2004.
Tutorial 10 Iterative Methods and Matrix Norms. 2 In an iterative process, the k+1 step is defined via: Iterative processes Eigenvector decomposition.
Compressed Sensing Compressive Sampling
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
KKT Practice and Second Order Conditions from Nash and Sofer
CHAPTER FIVE Orthogonality Why orthogonal? Least square problem Accuracy of Numerical computation.
+ Review of Linear Algebra Optimization 1/14/10 Recitation Sivaraman Balakrishnan.
The method of moments in dynamic optimization
Cs: compressed sensing
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Serge Andrianov Theory of Symplectic Formalism for Spin-Orbit Tracking Institute for Nuclear Physics Forschungszentrum Juelich Saint-Petersburg State University,
Machine Learning Weak 4 Lecture 2. Hand in Data It is online Only around 6000 images!!! Deadline is one week. Next Thursday lecture will be only one hour.
E XACT MATRIX C OMPLETION VIA CONVEX OPTIMIZATION E MMANUEL J. C ANDES AND B ENJAMIN R ECHT M AY 2008 Presenter: Shujie Hou January, 28 th,2011 Department.
Review of Linear Algebra Optimization 1/16/08 Recitation Joseph Bradley.
+ Quadratic Programming and Duality Sivaraman Balakrishnan.
Large-Scale Matrix Factorization with Missing Data under Additional Constraints Kaushik Mitra University of Maryland, College Park, MD Sameer Sheoreyy.
© Eric CMU, Machine Learning Support Vector Machines Eric Xing Lecture 4, August 12, 2010 Reading:
Inequalities for Stochastic Linear Programming Problems By Albert Madansky Presented by Kevin Byrnes.
Linear Programming Chapter 9. Interior Point Methods  Three major variants  Affine scaling algorithm - easy concept, good performance  Potential.
Inequality Constraints Lecture 7. Inequality Contraints (I) n A Review of Lagrange Multipliers –As we discussed last time, the first order necessary conditions.
Support Vector Machine: An Introduction. (C) by Yu Hen Hu 2 Linear Hyper-plane Classifier For x in the side of o : w T x + b  0; d = +1; For.
Approximation Algorithms based on linear programming.
1 Support Vector Machines: Maximum Margin Classifiers Machine Learning and Pattern Recognition: September 23, 2010 Piotr Mirowski Based on slides by Sumit.
Lecture 16: Image alignment
Support Vector Machines (SVM)
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
The Duality Theorem Primal P: Maximize
12. Principles of Parameter Estimation
Part 4 Nonlinear Programming
René Vidal and Xiaodong Fan Center for Imaging Science
Proving that a Valid Inequality is Facet-defining
Nonnegative polynomials and applications to learning
Duality for linear programming.
Basic Algorithms Christina Gallner
Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization Presenter: Xia Li.
Chap 9. General LP problems: Duality and Infeasibility
CS5321 Numerical Optimization
CSCI B609: “Foundations of Data Science”
Bounds for Optimal Compressed Sensing Matrices
1.3 Vector Equations.
Part 4 Nonlinear Programming
Rank-Sparsity Incoherence for Matrix Decomposition
I.4 Polyhedral Theory (NW)
I.4 Polyhedral Theory.
Proving that a Valid Inequality is Facet-defining
CIS 700: “algorithms for Big Data”
(Convex) Cones Def: closed under nonnegative linear combinations, i.e.
CS5321 Numerical Optimization
12. Principles of Parameter Estimation
Outline Sparse Reconstruction RIP Condition
Presentation transcript:

Nuclear Norm Heuristic for Rank Minimization Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization Necessary and Sufficient Conditions for Success of the Nuclear Norm Heuristic for Rank Minimization Presenter: Zhen Hu

Outline Introduction From Compressed Sensing to Rank Minimization Restricted Isometry and Recovery of Low-Rank Matrices Algorithms for Nuclear Norm Minimization Necessary and Sufficient Conditions for Success of the Nuclear Norm Heuristic for Rank Minimization Discussion and Future Developments 11/21/2018

Introduction An affine rank minimization problem Minimization of the l1 norm is a well known heuristic for the cardinality minimization problem. L1 heuristic can be a priori guaranteed to yield the optimal solution. The results from the compressed sensing literature might be extended to provide guarantees about the nuclear norm heuristic for the more general rank minimization problem. 11/21/2018

From Compressed Sensing 11/21/2018

Matrix Norm Frobenius Norm Operator Norm Nuclear Norm 11/21/2018

Convex Envelopes of Rank and Cardinality Functions 11/21/2018

Additive of Rank and Nuclear Norm 11/21/2018

Nuclear Norm Minimization This problem admits the primal-dual convex formulation The primal-dual pair of semidefinite programs 11/21/2018

Restricted Isometry and Recovery of Low Rank Matrices 11/21/2018

Nearly Isometric Families 11/21/2018

11/21/2018

Main Results 11/21/2018

Main Results 11/21/2018

Algorithms for Nuclear Norm Minimization The trade-offs between computational speed and guarantees on the accuracy of the resulting solution. Algorithms for Nuclear Norm Minimization Interior Point Methods for Semidefinite Programming For small problems where a high-degree of numerical precision is required, interior point methods for semidefinite programming can be directly applied to solve affine nuclear minimization problems. Projected Subgradient Methods Low-rank Parametrization SDPLR and the Method of Multipliers 11/21/2018

Numerical Experiments 11/21/2018

11/21/2018

Necessary and Sufficient Conditions The authors present a necessary and sufficient condition for the solution of the nuclear norm heuristic to coincide with the minimum rank solution in an affine space. The condition characterizes a particular property of the null space of the linear map which defines the affine space. The authors also present a reduction of the standard Linear Matrix Inequality (LMI) constrained rank minimization problem to a rank minimization problem with only equality constraints. 11/21/2018

Moreover, the authors show that when the linear map defining the constraint set is generated by sampling its entries independently from a Gaussian distribution, the null-space characterization holds with overwhelming probability provided the dimensions of the equality constraints are of appropriate size. The authors provide numerical experiments demonstrating that even when matrix dimensions are small, the nuclear norm heuristic does indeed always recover the minimum rank solution when the number of constraints is sufficiently large. 11/21/2018

Necessary and Sufficient Conditions The main optimization problem under study is The main concern is when the optimal solution coincides with the optimal solution of 11/21/2018

11/21/2018

Theorem 3.1 The following theorem generalizes this null-space criterion to a critical property that guarantees when the nuclear norm heuristic finds the minimum rank solution of A(X)=b for all values of the vector b. A particular property of the null space of the linear map 11/21/2018

Lemma 4.1 For the proof of Theorem 3.1 11/21/2018

Theorem 3.2 Reduction to the affine case! 11/21/2018

Weak Bound Probabilistic generation of constraints satisfying null space characterization 11/21/2018

Strong Bound 11/21/2018

Performance 11/21/2018

Discussion Having illustrated the natural connections between affine rank minimization and affine cardinality minimization, we were able to draw on these parallels to determine scenarios where the nuclear norm heuristic was able to exactly solve the rank minimization problem. Future developments Factored measurements and alternative ensembles Noisy measurements and low rank approximation Incoherent ensembles and partially observed transforms Alternative numerical methods Geometric interpretations Parsimonious models and optimization 11/21/2018

Question? 11/21/2018