Presentation is loading. Please wait.

Presentation is loading. Please wait.

Learning Measurement Matrices for Redundant Dictionaries Richard Baraniuk Rice University Chinmay Hegde MIT Aswin Sankaranarayanan CMU.

Similar presentations

Presentation on theme: "Learning Measurement Matrices for Redundant Dictionaries Richard Baraniuk Rice University Chinmay Hegde MIT Aswin Sankaranarayanan CMU."— Presentation transcript:

1 Learning Measurement Matrices for Redundant Dictionaries Richard Baraniuk Rice University Chinmay Hegde MIT Aswin Sankaranarayanan CMU

2 Sparse Recovery Sparsity rocks, etc. Previous talk focused mainly on signal inference (ex: classification, NN search) This talk focuses on signal recovery

3 Compressive Sensing Sensing via randomized dimensionality reduction random measurements sparse signal nonzero entries Recovery:solve an ill-posed inverse problem exploit the geometrical structure of sparse/compressible signals

4 Gaussian measurements incoherent with any fixed orthonormal basis (with high probability) Ex: frequency domain: General Sparsifying Bases

5 Sparse Modeling: Approach 1 Step 1: Choose a signal model with structure –e.g. bandlimited, smooth with r vanishing moments, etc. Step 2: Analytically design a sparsifying basis/frame that exploits this structure –e.g. DCT, wavelets, Gabor, etc. DCT Wavelets Gabor ?

6 Sparse Modeling: Approach 2 Learn the sparsifying basis/frame from training data Problem formulation: given a large number of training signals, design a dictionary D that simultaneously sparsifies the training data Called sparse coding / dictionary learning

7 Dictionaries Dictionary: an NxQ matrix whose columns are used as basis functions for the data Convention: assume columns are unit-norm More columns than rows, so dictionary is redundant / overcomplete

8 Dictionary Learning Rich vein of theoretical and algorithmic work Olshausen and Field [‘97], Lewicki and Sejnowski [’00], Elad [‘06], Sapiro [‘08] Typical formulation: Given training data Solve: Several efficient algorithms, ex: K-SVD

9 Dictionary Learning Successfully applied to denoising, deblurring, inpainting, demosaicking, super-resolution, … –State-of-the-art results in many of these problems Aharon and Elad ‘06

10 Dictionary Coherence Suppose that the learned dictionary is normalized to have unit -norm columns: The mutual coherence of D is defined as Geometrically, represents the cosine of the minimum angle between the columns of D, smaller is better Crucial parameter in analysis as well as practice (line of work starting with Tropp [04])

11 Dictionaries and CS Can extend CS to work with non-orthonormal, redundant dictionaries Coherence of determines recovery success Rauhut et al. [08], Candes et al. [10] Fortunately, random guarantees low coherence Holographic basis

12 Geometric Intuition Columns of D: points on the unit sphere Coherence: minimum angle between the vectors J-L Lemma: Random projections approximately preserve angles between vectors

13 Q: Can we do better than random projections for dictionary-based CS? Q restated: For a given dictionary D, find the best CS measurement matrix

14 Optimization Approach Assume that a good dictionary D has been provided. Goal: Learn the best for this particular D As before, want the “shortest” matrix such that the coherence of is at most some parameter To avoid degeneracies caused by a simple scaling, also want that does not shrink columns much:

15 A NuMax-like Framework Convert quadratic constraints in into linear constraints in (via the “lifting trick”) Use a nuclear-norm relaxation of the rank Simplified problem:

16 Alternating Direction Method of Multipliers (ADMM) - solve for P using spectral thresholding - solve for L using least-squares - solve for q using “squishing” Convergence rate depends on the size of the dictionary (since #constraints = ) Algorithm: “NuMax-Dict” [HSYB12]

17 NuMax vs. NuMax-Dict Same intuition, trick, algorithm, etc; Key enabler is that coherence is intrinsically a quadratic function of the data Key difference: the (linearized) constraints are no longer symmetric –We have constraints of the form –This might result in intermediate P estimates having complex eigenvalues, so the notion of spectral thresholding needs to be slightly modified

18 Experimental Results

19 Expt 1: Synthetic Dictionary Generic dictionary: random w/ unit norm. columns Dictionary size: 64x128 We construct different measurement matrices: Random NuMax-Dict Algorithm by Elad [06] Algorithm by Duarte-Carvajalino & Sapiro [08] We generate K=3 sparse signals with Gaussian amplitudes, add 30dB measurement noise Recovery using OMP Measure recovery SNR, plot as a function of M

20 Exp 1: Synthetic Dictionary

21 Expt 2: Practical Dictionaries 2x overcomplete DCT dictionary, same parameters 2x overcomplete dictionary learned on 8x8 patches of a real-world image (Barbara) using K-SVD Recovery using OMP

22 Analysis Exact problem seems to be hard to analyze But, as in NuMax, can provide analytical bounds in the special case where the measurement matrix is further constrained to be orthonormal

23 Orthogonal Sensing of Dictionary-Sparse Signals Given a dictionary D, find the orthonormal measurement matrix that provides the best possible coherence From a geometric perspective, ortho-projections cannot improve coherence, so necessarily

24 Semidefinite Relaxation The usual trick: Lifting and trace-norm relaxation

25 Theoretical Result Theorem: For any given redundant dictionary D, denote its mutual coherence by. Denote the optimum of the (nonconvex) problem as Then, there exists a method to produce a rank-2M ortho matrix such that the coherence of is at most i.e., We can obtain close to optimal performance, but pay a price of a factor 2 in the number of measurements

26 Conclusions NuMax-Dict performance comparable to the best existing algorithms Principled convex optimization framework Efficient ADMM-type algorithm that exploits the rank-1 structure of the problem Upshot: possible to incorporate other structure into the measurement matrix, such as positivity, sparsity, etc.

27 Open Question Above framework assumes a two-step approach: first construct a redundant dictionary (analytically or from data) and then construct a measurement matrix Given a large number of training data, how to efficiently solve jointly for both the dictionary and the sensing matrix? (Approach introduced in DC-Sapiro [08])

Download ppt "Learning Measurement Matrices for Redundant Dictionaries Richard Baraniuk Rice University Chinmay Hegde MIT Aswin Sankaranarayanan CMU."

Similar presentations

Ads by Google