AMSC 6631 Sparse Solutions of Linear Systems of Equations and Sparse Modeling of Signals and Images: Midyear Report Alfredo Nava-Tudela John J. Benedetto,

Slides:



Advertisements
Similar presentations
The simplex algorithm The simplex algorithm is the classical method for solving linear programs. Its running time is not polynomial in the worst case.
Advertisements

Shortest Vector In A Lattice is NP-Hard to approximate
Ch 7.7: Fundamental Matrices
Compressive Sensing IT530, Lecture Notes.
Multi-Label Prediction via Compressed Sensing By Daniel Hsu, Sham M. Kakade, John Langford, Tong Zhang (NIPS 2009) Presented by: Lingbo Li ECE, Duke University.
MMSE Estimation for Sparse Representation Modeling
Joint work with Irad Yavneh
Pattern Recognition and Machine Learning
Online Performance Guarantees for Sparse Recovery Raja Giryes ICASSP 2011 Volkan Cevher.
Wangmeng Zuo, Deyu Meng, Lei Zhang, Xiangchu Feng, David Zhang
MATH 685/ CSI 700/ OR 682 Lecture Notes
CSCI 317 Mike Heroux1 Sparse Matrix Computations CSCI 317 Mike Heroux.
Chapter 5 Orthogonality
Sparse and Overcomplete Data Representation
Math for CSLecture 41 Linear Least Squares Problem Over-determined systems Minimization problem: Least squares norm Normal Equations Singular Value Decomposition.
Optimized Projection Directions for Compressed Sensing Michael Elad The Computer Science Department The Technion – Israel Institute of technology Haifa.
Information Retrieval in Text Part III Reference: Michael W. Berry and Murray Browne. Understanding Search Engines: Mathematical Modeling and Text Retrieval.
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Matrix sparsification and the sparse null space problem Lee-Ad GottliebWeizmann Institute Tyler NeylonBynomial Inc. TexPoint fonts used in EMF. Read the.
Computing Sketches of Matrices Efficiently & (Privacy Preserving) Data Mining Petros Drineas Rensselaer Polytechnic Institute (joint.
Recent Trends in Signal Representations and Their Role in Image Processing Michael Elad The CS Department The Technion – Israel Institute of technology.
Tutorial 10 Iterative Methods and Matrix Norms. 2 In an iterative process, the k+1 step is defined via: Iterative processes Eigenvector decomposition.
Linear Least Squares QR Factorization. Systems of linear equations Problem to solve: M x = b Given M x = b : Is there a solution? Is the solution unique?
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
A Weighted Average of Sparse Several Representations is Better than the Sparsest One Alone Michael Elad The Computer Science Department The Technion –
A Sparse Solution of is Necessarily Unique !! Alfred M. Bruckstein, Michael Elad & Michael Zibulevsky The Computer Science Department The Technion – Israel.
Topics in MMSE Estimation for Sparse Approximation Michael Elad The Computer Science Department The Technion – Israel Institute of technology Haifa 32000,
Alfredo Nava-Tudela John J. Benedetto, advisor
Compressed Sensing Compressive Sampling
1 Information Retrieval through Various Approximate Matrix Decompositions Kathryn Linehan Advisor: Dr. Dianne O’Leary.
Cs: compressed sensing
Simplex method (algebraic interpretation)
SVD: Singular Value Decomposition
Orthogonality and Least Squares
Orthogonalization via Deflation By Achiya Dax Hydrological Service Jerusalem, Israel
Elementary Linear Algebra Anton & Rorres, 9th Edition
MATH 685/ CSI 700/ OR 682 Lecture Notes Lecture 4. Least squares.
Section 2.3 Properties of Solution Sets
A Note on Rectangular Quotients By Achiya Dax Hydrological Service Jerusalem, Israel
Sparse Signals Reconstruction Via Adaptive Iterative Greedy Algorithm Ahmed Aziz, Ahmed Salim, Walid Osamy Presenter : 張庭豪 International Journal of Computer.
Biointelligence Laboratory, Seoul National University
Elementary Linear Algebra Anton & Rorres, 9 th Edition Lecture Set – 07 Chapter 7: Eigenvalues, Eigenvectors.
Sparse & Redundant Representation Modeling of Images Problem Solving Session 1: Greedy Pursuit Algorithms By: Matan Protter Sparse & Redundant Representation.
A Weighted Average of Sparse Representations is Better than the Sparsest One Alone Michael Elad and Irad Yavneh SIAM Conference on Imaging Science ’08.
OR  Now, we look for other basic feasible solutions which gives better objective values than the current solution. Such solutions can be examined.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Compressive Sensing Techniques for Video Acquisition EE5359 Multimedia Processing December 8,2009 Madhu P. Krishnan.
Matrices CHAPTER 8.9 ~ Ch _2 Contents  8.9 Power of Matrices 8.9 Power of Matrices  8.10 Orthogonal Matrices 8.10 Orthogonal Matrices 
From Sparse Solutions of Systems of Equations to Sparse Modeling of Signals and Images Alfred M. Bruckstein (Technion), David L. Donoho (Stanford), Michael.
Basic Algorithms Christina Gallner
Chapter 10: Solving Linear Systems of Equations
Singular Value Decomposition
Quantum Two.
Sparse and Redundant Representations and Their Applications in
Sparse and Redundant Representations and Their Applications in
A Motivating Application: Sensor Array Signal Processing
Orthogonality and Least Squares
Singular Value Decomposition SVD
Sparse and Redundant Representations and Their Applications in
Optimal sparse representations in general overcomplete bases
Chapter 8. General LP Problems
Stability Analysis of Linear Systems
Recursively Adapted Radial Basis Function Networks and its Relationship to Resource Allocating Networks and Online Kernel Learning Weifeng Liu, Puskal.
Maths for Signals and Systems Linear Algebra in Engineering Lecture 6, Friday 21st October 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL.
Chapter 8. General LP Problems
CIS 700: “algorithms for Big Data”
Chapter 8. General LP Problems
Orthogonality and Least Squares
Outline Sparse Reconstruction RIP Condition
Presentation transcript:

AMSC 6631 Sparse Solutions of Linear Systems of Equations and Sparse Modeling of Signals and Images: Midyear Report Alfredo Nava-Tudela John J. Benedetto, advisor

AMSC 6632 Outline - Problem: Find “sparse solutions” of Ax = b. - Definitions of “sparse solution”. - How do we find sparse solutions? The Orthogonal Matching Pursuit (OMP) - Some theoretical results. - Implementation and validation, some details. - Validation results. - Conclusions/Recapitulation. - Project timeline, current status. - References.

AMSC 6633 Problem Let A be an n by m matrix, with n < m, and rank(A) = n. We want to solve Ax = b, where b is a data or signal vector, and x is the solution with the fewest number of non-zero entries possible, that is, the “sparsest” one. Observations: - A is underdetermined and, since rank(A) = n, there is an infinite number of solutions. Good! - How do we find the “sparsest” solution? What does this mean in practice? Is there a unique sparsest solution?

AMSC 6634 Why is this problem relevant? “Sparsity” equals compression: Assume Ax = b. If x is sparse, and b is dense, store x! Image compression techniques, such as JPEG [6] or JPEG-2000 [5], are based in this idea, where a linear transformation provides a sparse representation within an error margin of the original image. In the second half of this project, we shall further explore the role of sparsity in signal compression.

AMSC 6635 Definitions of “sparse” - Convenient to introduce the l 0 “norm” [1]: ||x|| 0 = # {k : x k ≠ 0} - (P 0 ): min x ||x|| 0 subject to ||Ax - b|| 2 = 0 - (P 0  ): min x ||x|| 0 subject to ||Ax - b|| 2 <  Observations: In practice, (P 0  ) is the working definition of sparsity as it is the only one that is computationally practical. Solving (P 0  ) is NP-hard [2].

AMSC 6636 Finding sparse solutions: OMP Orthogonal Matching Pursuit algorithm [1]:

AMSC 6637 Some theoretical results [1] Definition: The mutual coherence of a matrix A is the number Theorem: If x solves Ax = b, and ||x|| 0 < (1+  (A) -1 )/2, then x is the sparsest solution. That is, if y ≠ x also solves the equation, then ||x|| 0 < ||y|| 0. Theorem: For a system of linear equations Ax = b (A an n by m matrix, n < m, and rank(A) = n), if a solution x exists obeying ||x|| 0 < (1+  (A) -1 )/2, then an OMP run with threshold parameter e 0 = 0 is guaranteed to find x exactly.

AMSC 6638 Implementation and Validation In light of these theoretical results, we can envision the following roadmap to validate an implementation of OMP. - We have a simple theoretical criterion to guarantee both solution uniqueness and OMP convergence: If x is a solution to Ax = b, and ||x|| 0 < (1+  (A) -1 )/2, then x is the unique sparsest solution to Ax = b and OMP will find it. - Hence, given a full-rank n by m matrix A (n < m), compute  (A), and find the largest integer k smaller than or equal to (1+  (A) -1 )/2. That is, k = floor((1+  (A) -1 )/2).

AMSC 6639 Implementation and Validation - Build a vector x with exactly k non-zero entries and produce a right hand side vector b = Ax. This way, you have a known sparsest solution x to which to compare the output of any OMP implementation. - Pass A, b, and  0 to OMP to produce a solution vector x omp = OMP(A,b,  0 ). - If OMP terminates after k iterations and ||Ax omp - b|| <  0, for all possible x and  0 > 0, then the OMP implementation would have been validated. Caveat: The theoretical proofs assume infinite precision.

AMSC Implementation and Validation - Some implementation details worth discussing: The core of the algorithm is found in the following three steps. We will discuss in detail our implementation of the “Update Support” and “Update Provisional Solution” steps.

AMSC Implementation and Validation ---Initialization--- k = 0; activeCol = []; % will contain the indices of the active columns of A. epsilon = zeros(m,1); % contains the errors epsilon(j) described above. ---Inside Main Loop--- k = k + 1; % Sweep for j = 1:m a_j = A(:,j); z_j = a_j'*r0/norm(a_j)^2; epsilon(j) = norm(z_j*a_j - r0)^2; end % Update Support maxValueEpsilon = max(epsilon); epsilon(activeCol) = maxValueEpsilon; [minValueEpsilon, j_0] = min(epsilon); % j_0 is the new index to add. activeCol(k) = j_0; % update the set of active columns of A.

AMSC Implementation and Validation A 3 = QR = Q 1 R 1 + Q 2 R 2 = Q 1 R = Q 1 R 1 Solve the linear system A 3 x * = b, with x *  R 3. We have: A 3 x * = QRx * = Q 1 R 1 x * = b  Q 1 T Q 1 R 1 x * = Q 1 T b (1) See [3] for more on the QR decomposition.

AMSC Implementation and Validation (1):Q 1 T Q 1 R 1 x * = Q 1 T b  R 1 x * = Q 1 T b  x * = (R 1 ) -1 Q 1 T b, Observation: where we can obtain the last equation because A is a full rank matrix, and therefore A 3 is too, implying (R 1 ) -1 exists.

AMSC Implementation and Validation The minimizer x k=3 of ||Ax - b|| 2 2, subject to support{x} = S k=3, is then obtained when we solve A 3 x * = b, with x *  R 3, and we set x k=3 equal to the “natural embedding” of x * into the zero vector 0  R m. ---Initialization--- x0 = zeros(m,1); ---Inside Main Loop--- % Update the provisional solution by solving an equivalent unconstrained % least squares problem. A_k = A(:,activeCol); [Q,R] = qr(A_k); x0(activeCol) = R(1:k,:) \ Q(:,1:k)'*b;

AMSC Validation Results We ran two experiments: 1)A  R 100x200, with entries in N(0,1) i.i.d. for which  (A) = , corresponding to k = 1  . 2)A  R 200x400, with entries in N(0,1) i.i.d. for which  (A) = , corresponding to k = 2  . Observations: - A will be full-rank with probability 1 [1]. - For full-rank matrices A of size n x m, the mutual coherence satisfies  (A)   {(m - n)/(n  (m - 1))} [5]. That is, the upper bound of  = 0.5*(1 + 1/  (A)) can be made as big as needed, provided n and m are big enough.

AMSC Validation Results For each matrix A, we chose 100 vectors with k non-zero entries whose positions were chosen at random, and whose entries were in N(0,1). Then, for each such vector x, we built a corresponding right hand side vector b = Ax. Each of these vectors would then be the unique sparsest solution to Ax = b, and OMP should be able to find them. Finally, given  0 > 0, if our implementation of OMP were correct, it should stop after k steps (or less), and if x OMP = OMP(A,b,  0 ), then ||b - Ax OMP || <  0.

AMSC Validation Results k = 1

AMSC Validation Results k = 1

AMSC Validation Results k = 1

AMSC Validation Results k = 1

AMSC Validation Results k = 1

AMSC Validation Results k = 2

AMSC Validation Results k = 2

AMSC Validation Results k = 2

AMSC Validation Results k = 2

AMSC Validation Results k = 2

AMSC Conclusions/Recapitulation - There are simple criteria to test the uniqueness of a given sparse solution. - There are algorithms that find sparse solutions, e.g., OMP; and their convergence can be guaranteed when there are “sufficiently sparse” solutions. - Our implementation of OMP is successful up to machine precision as predicted by current theoretical bounds.

AMSC Project timeline - Oct. 15, 2010: Complete project proposal.  - Oct Nov. 5: Implement OMP.  - Nov. 8 - Nov. 26: Validate OMP.  - Nov Dec. 3: Write mid-year report. 1/2 - Dec. 6 - Dec. 10: Prepare mid-year oral presentation.  - Some time after that, give mid-year oral presentation.  - Jan Feb. 11: Testing of OMP. Reproduce paper results. - Jan Apr. 8: Testing of OMP on A = [DCT,DWT]. - Apr Apr. 22: Write final report. - Apr Apr. 29: Prepare final oral presentation. - Some time after that, give final oral presentation.

AMSC References [1] A. M. Bruckstein, D. L. Donoho, and M. Elad, From sparse solutions of systems of equations to sparse modeling of signals and images, SIAM Review, 51 (2009), pp. 34–81. [2] B. K. Natarajan, Sparse approximate solutions to linear systems, SIAM Journal on Computing, 24 (1995), pp [3] G. W. Stewart, Introduction to Matrix Computations, Academic Press, [4] T. Strohmer and R. W. Heath, Grassmanian frames with applications to coding and communication, Appl. Comput. Harmon. Anal., 14 (2004), pp [5] D. S. Taubman and M. W. Mercellin, JPEG 2000: Image Compression Funda- mentals, Standards and Practice, Kluwer Academic Publishers, [6] G. K. Wallace, The JPEG still picture compression standard, Communications of the ACM, 34 (1991), pp