Combinatorial Selection and Least Absolute Shrinkage via The CLASH Operator Volkan Cevher Laboratory for Information and Inference Systems – LIONS / EPFL.

Slides:



Advertisements
Similar presentations
An Introduction to Compressed Sensing Student : Shenghan TSAI Advisor : Hsuan-Jung Su and Pin-Hsun Lin Date : May 02,
Advertisements

On the Power of Adaptivity in Sparse Recovery Piotr Indyk MIT Joint work with Eric Price and David Woodruff, 2011.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Joint work with Irad Yavneh
Pixel Recovery via Minimization in the Wavelet Domain Ivan W. Selesnick, Richard Van Slyke, and Onur G. Guleryuz *: Polytechnic University, Brooklyn, NY.
Sparse Euclidean projections onto convex sets Volkan Cevher Laboratory for Information and Inference Systems – LIONS / EPFL joint.
Short-course Compressive Sensing of Videos Venue CVPR 2012, Providence, RI, USA June 16, 2012 Organizers: Richard G. Baraniuk Mohit Gupta Aswin C. Sankaranarayanan.
Learning Measurement Matrices for Redundant Dictionaries Richard Baraniuk Rice University Chinmay Hegde MIT Aswin Sankaranarayanan CMU.
Online Performance Guarantees for Sparse Recovery Raja Giryes ICASSP 2011 Volkan Cevher.
Submodular Dictionary Selection for Sparse Representation Volkan Cevher Laboratory for Information and Inference Systems - LIONS.
Multi-Task Compressive Sensing with Dirichlet Process Priors Yuting Qi 1, Dehong Liu 1, David Dunson 2, and Lawrence Carin 1 1 Department of Electrical.
Beyond Nyquist: Compressed Sensing of Analog Signals
The Analysis (Co-)Sparse Model Origin, Definition, and Pursuit
Wangmeng Zuo, Deyu Meng, Lei Zhang, Xiangchu Feng, David Zhang
Variational Inference in Bayesian Submodular Models
Approximate Message Passing for Bilinear Models
Extensions of wavelets
More MR Fingerprinting
Learning With Dynamic Group Sparsity Junzhou Huang Xiaolei Huang Dimitris Metaxas Rutgers University Lehigh University Rutgers University.
ECE Department Rice University dsp.rice.edu/cs Measurements and Bits: Compressed Sensing meets Information Theory Shriram Sarvotham Dror Baron Richard.
Sparse Representation and Compressed Sensing: Theory and Algorithms
“Random Projections on Smooth Manifolds” -A short summary
Volkan Cevher, Marco F. Duarte, and Richard G. Baraniuk European Signal Processing Conference 2008.
Quantization and compressed sensing Dmitri Minkin.
Richard Baraniuk Rice University dsp.rice.edu/cs Compressive Signal Processing.
Random Convolution in Compressive Sampling Michael Fleyer.
Introduction to Compressive Sensing
Rice University dsp.rice.edu/cs Distributed Compressive Sensing A Framework for Integrated Sensing and Processing for Signal Ensembles Marco Duarte Shriram.
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
Hybrid Dense/Sparse Matrices in Compressed Sensing Reconstruction
Compressed Sensing Compressive Sampling
Model-based Compressive Sensing
An ALPS’ view of Sparse Recovery Volkan Cevher Laboratory for Information and Inference Systems - LIONS
ACCESS IC LAB Graduate Institute of Electronics Engineering, NTU Reconstruction Algorithms for Compressive Sensing II Presenter: 黃乃珊 Advisor: 吳安宇 教授 Date:
Richard Baraniuk Rice University Model-based Sparsity.
Compressive Sampling: A Brief Overview
Sensing via Dimensionality Reduction Structured Sparsity Models Volkan Cevher
PATTERN RECOGNITION AND MACHINE LEARNING
IC Research Seminar A Teaser Prof. Dr. Volkan Cevher LIONS/Laboratory for Information and Inference Systems
Game Theory Meets Compressed Sensing
Compressive Sensing A New Approach to Signal Acquisition and Processing Richard Baraniuk Rice University.
Richard Baraniuk Chinmay Hegde Marco Duarte Mark Davenport Rice University Michael Wakin University of Michigan Compressive Learning and Inference.
Recovery of Clustered Sparse Signals from Compressive Measurements
Cs: compressed sensing
Introduction to Compressive Sensing
Recovering low rank and sparse matrices from compressive measurements Aswin C Sankaranarayanan Rice University Richard G. Baraniuk Andrew E. Waters.
EE369C Final Project: Accelerated Flip Angle Sequences Jan 9, 2012 Jason Su.
Learning With Structured Sparsity
Source Localization on a budget Volkan Cevher Rice University Petros RichAnna Martin Lance.
Model-Based Compressive Sensing Presenter: Jason David Bonior ECE / CMR Tennessee Technological University November 5, 2010 Reading Group (Richard G. Baraniuk,
SCALE Speech Communication with Adaptive LEarning Computational Methods for Structured Sparse Component Analysis of Convolutive Speech Mixtures Volkan.
Compressible priors for high-dimensional statistics Volkan Cevher LIONS/Laboratory for Information and Inference Systems
Shriram Sarvotham Dror Baron Richard Baraniuk ECE Department Rice University dsp.rice.edu/cs Sudocodes Fast measurement and reconstruction of sparse signals.
Zhilin Zhang, Bhaskar D. Rao University of California, San Diego March 28,
From Sparse Solutions of Systems of Equations to Sparse Modeling of Signals and Images Alfred M. Bruckstein (Technion), David L. Donoho (Stanford), Michael.
n-pixel Hubble image (cropped)
Jeremy Watt and Aggelos Katsaggelos Northwestern University
Lecture 15 Sparse Recovery Using Sparse Matrices
Learning With Dynamic Group Sparsity
Nonnegative polynomials and applications to learning
Basic Algorithms Christina Gallner
Probabilistic Models for Linear Regression
Nuclear Norm Heuristic for Rank Minimization
Bounds for Optimal Compressed Sensing Matrices
Sudocodes Fast measurement and reconstruction of sparse signals
Optimal sparse representations in general overcomplete bases
CIS 700: “algorithms for Big Data”
Sudocodes Fast measurement and reconstruction of sparse signals
Outline Sparse Reconstruction RIP Condition
Subspace Expanders and Low Rank Matrix Recovery
Presentation transcript:

Combinatorial Selection and Least Absolute Shrinkage via The CLASH Operator Volkan Cevher Laboratory for Information and Inference Systems – LIONS / EPFL & Idiap Research Institute joint work with my PhD student Anastasios EPFL

Linear Inverse Problems Machine learningdictionary of features Compressive sensingnon-adaptive measurements Information theorycoding frame Theoretical computer science sketching matrix / expander

Linear Inverse Problems Challenge:

Approaches Deterministic Probabilistic Prior sparsity compressibility Metric likelihood function

A Deterministic View Model-based CS (circa Aug 2008)

1.Sparse or compressible not sufficient alone 2.Projection information preserving (stable embedding / special null space) 3.Decoding algorithms tractable My Insights on Compressive Sensing

Sparse signal: only K out of N coordinates nonzero –model: union of all K -dimensional subspaces aligned w/ coordinate axes Example: 2-sparse in 3-dimensions Signal Priors support:

Sparse signal:only K out of N coordinates nonzero –model: union of all K -dimensional subspaces aligned w/ coordinate axes Structured sparse signal: reduced set of subspaces (or model-sparse) –model:a particular union of subspaces ex: clustered or dispersed sparse patterns sorted index Signal Priors

Sparse signal:only K out of N coordinates nonzero –model: union of all K -dimensional subspaces aligned w/ coordinate axes Structured sparse signal: reduced set of subspaces (or model-sparse) –model:a particular union of subspaces ex: clustered or dispersed sparse patterns sorted index Signal Priors

Sparse Recovery Algorithms Goal:given recover and convex optimization formulations –basis pursuit, Lasso, BP denoising… –iterative re-weighted algorithms Hard thresholding algorithms: ALPS, CoSaMP, SP,… Greedy algorithms: OMP, MP,…

GeometricCombinatorialProbabilistic Encodingatomic norm / convex relaxation non-convex union-of-subspaces compressible / sparse priors Example Algorithm Basis pursuit, Lasso, basis pursuit denoising… IHT, CoSaMP, SP, ALPS, OMP… Variational Bayes, EP, Approximate message passing (AMP)… Sparse Recovery Algorithms

GeometricCombinatorialProbabilistic Encodingatomic norm / convex relaxation non-convex union-of-subspaces compressible / sparse priors Example Algorithm Basis pursuit, Lasso, basis pursuit denoising… IHT, CoSaMP, SP, ALPS, OMP… Variational Bayes, EP, Approximate message passing (AMP)… The Clash Operator Sparse Recovery Algorithms

A Tale of Two Algorithms Soft thresholding

A Tale of Two Algorithms Soft thresholding Bregman distance Structure in optimization:

A Tale of Two Algorithms Soft thresholding ALGO: cf. J. Duchi et al. ICML 2008 Bregman distance majorization-minimization

A Tale of Two Algorithms Soft thresholding Bregman distance slower

A Tale of Two Algorithms Soft thresholding Is x* what we are looking for? local “unverifiable” assumptions: –ERC/URC condition –compatibility condition … (local  global / dual certification / random signal models)

A Tale of Two Algorithms Hard thresholding ALGO: sort and pick the largest K

A Tale of Two Algorithms Hard thresholding percolations What could possibly go wrong with this naïve approach?

A Tale of Two Algorithms Hard thresholding we can tiptoe among percolations! another variant has Global “unverifiable” assumption: GraDes:

Model: K-sparse coefficients RIP: stable embedding Restricted Isometry Property K-planes IID sub-Gaussian matrices Remark: implies convergence of convex relaxations also e.g.,

A Model-based CS Algorithm Model-based hard thresholding Global “unverifiable” assumption:

Tree-Sparse Model: K-sparse coefficients +significant coefficients lie on a rooted subtree Sparse approx: find best set of coefficients –sorting –hard thresholding Tree-sparse approx: find best rooted subtree of coefficients –condensing sort and select [Baraniuk] –dynamic programming [Donoho]

Model: K-sparse coefficients RIP: stable embedding Sparse K-planes IID sub-Gaussian matrices

Tree-Sparse Model: K-sparse coefficients +significant coefficients lie on a rooted subtree Tree-RIP: stable embedding K-planes IID sub-Gaussian matrices

Tree-Sparse Signal Recovery target signal CoSaMP, (MSE=1.12) L1-minimization (MSE=0.751) Tree-sparse CoSaMP (MSE=0.037) N=1024 M=80

Tree-Sparse Signal Recovery Number samples for correct recovery Piecewise cubic signals + wavelets Models/algorithms: –compressible (CoSaMP) –tree-compressible (tree-CoSaMP)

Model CS in Context Basis pursuit and Lasso exploit geometry<>interplay of arbitrary selection<>difficulty of interpretation cannot leverage further structure Structured-sparsity inducing norms “customize” geometry<>“mixing” of norms over groups / Lovasz extension of submodular set functions inexact selections Structured-sparsity via OMP / Model-CS greedy selection <>cannot leverage geometry exact selection<>cannot leverage geometry for selection

Model CS in Context Basis pursuit and Lasso exploit geometry<>interplay of arbitrary selection<>difficulty of interpretation cannot leverage further structure Structured-sparsity inducing norms “customize” geometry<>“mixing” of norms over groups / Lovasz extension of submodular set functions inexact selections Structured-sparsity via OMP / Model-CS greedy selection <>cannot leverage geometry exact selection<>cannot leverage geometry for selection Or, can it?

Enter CLASH

Importance of Geometry A subtle issue 2 solutions! Which one is correct?

Importance of Geometry A subtle issue 2 solutions! Which one is correct? EPIC FAIL

CLASH Pseudocode Algorithm –Active set expansion –Greedy descend –Combinatorial selection –Least absolute shrinkage –De-bias with convex constraint & Minimum 1-norm solution still makes sense!

CLASH Pseudocode Algorithm –Active set expansion –Greedy descend –Combinatorial selection –Least absolute shrinkage –De-bias with convex constraint

Geometry of CLASH Combinatorial selection + least absolute shrinkage &

Geometry of CLASH Combinatorial selection + least absolute shrinkage combinatorial origami &

Combinatorial Selection A different view of the model-CS workhorse (Lemma) support of the solution <> modular approximation problem where indexing set

PMAP An algorithmic generalization of union-of-subspaces Polynomial time modular epsilon-approximation property: Sets with PMAP-0 –Matroids uniform matroids<>regular sparsity partition matroids<>block sparsity (disjoint groups) cographic matroids<>rooted connected tree group adapted hull model –Totally unimodular systems mutual exclusivity<>neuronal spike model interval constraints<>sparsity within groups Model-CS is applicable for all these cases!

PMAP An algorithmic generalization of union-of-subspaces Polynomial time modular epsilon-approximation property: Sets with PMAP-epsilon –Knapsack multi-knapsack constraints weighted multi-knapsack quadratic knapsack (?) –Define algorithmically! … selector variables

PMAP An algorithmic generalization of union-of-subspaces Polynomial time modular epsilon-approximation property: Sets with PMAP-epsilon –Knapsack –Define algorithmically! Sets with PMAP-??? –pairwise overlapping groups<>mincut with cardinality constraint

CLASH Approximation Guarantees PMAP / downward compatibility –precise formulae are in the paper Isometry requirement (PMAP-0)<>

Examples Model: (K,C)-clustered model O(KCN) – per iteration ~10-15 iterations Model: partition model / TU LP – per iteration ~20-25 iterations sparse matrix

Examples CCD array readout via noiselets

Conclusions CLASH<>combinatorial selection + convex geometry   model-CS PMAP-epsilon<>inherent difficulty in combinatorial selection –beyond simple selection towards provable solution quality + runtime/space bounds –algorithmic definition of sparsity + many models matroids, TU, knapsack,… Postdoc LIONS / EPFL contact:

Postdoc LIONS / EPFL contact:

Sparse signal: only K out of N coordinates nonzero –model: union of K -dimensional subspaces Compressible signal: sorted coordinates decay rapidly to zero well-approximated by a K -sparse signal (simply by thresholding) sorted index Signal Priors

Compressible Signals Real-world signals are compressible, not sparse Recall: compressible <> well approximated by sparse –compressible signals lie close to a union of subspaces –ie: approximation error decays rapidly as If has RIP, then both sparse and compressible signals are stably recoverable sorted index

Model-Compressible Signals Model-compressible <> well approximated by model-sparse –model-compressible signals lie close to a reduced union of subspaces –ie: model-approx error decays rapidly as

Model-Compressible Signals Model-compressible <> well approximated by model-sparse –model-compressible signals lie close to a reduced union of subspaces –ie: model-approx error decays rapidly as While model-RIP enables stable model-sparse recovery, model-RIP is not sufficient for stable model-compressible recovery at !

Stable Recovery Stable model-compressible signal recovery at requires that have both: –RIP + Restricted Amplification Property RAmP: controls nonisometry of in the approximation’s residual subspaces optimal K-term model recovery (error controlled by RIP) optimal 2K-term model recovery (error controlled by RIP) residual subspace (error not controlled by RIP)

Tree-RIP, Tree-RAmP Theorem: An MxN iid subgaussian random matrix has the Tree(K)-RIP if Theorem: An MxN iid subgaussian random matrix has the Tree(K)-RAmP if

Other Useful Models When the model-based framework makes sense: –model with  exact projection algorithm –sensing matrix with  model-RIP  model-RAmP Ex: block sparsity / signal ensembles [Tropp, Gilbert, Strauss], [Stojnic, Parvaresh, Hassibi], [Eldar, Mishali], [Baron, Duarte et al], [Baraniuk, VC, Duarte, Hegde] Ex: clustered signals / background subtraction [VC, Duarte, Hegde, Baraniuk], [VC, Indyk, Hegde, Baraniuk] Ex: neuronal spike trains [Hegde, Duarte, VC – best student paper award]

Combinatorial Selection A different view of the model-CS workhorse (Lemma) support of the solution <> modular approximation problem where Sets endowed with an ALGO with two properties – –ALGO: (pseudo) polynomial time Denote the projection as indexing set

Performance of Recovery With model-based ALPS, CoSaMP, SP with model-RIP Model-sparse signals –noise-free measurements: exact recovery –noisy measurements: stable recovery Model-compressible signals (via RaMP) –recovery as good as K-model-sparse approximation CS recovery error signal K-term model approx error noise