GRADIENT PROJECTION FOR SPARSE RECONSTRUCTION: APPLICATION TO COMPRESSED SENSING AND OTHER INVERSE PROBLEMS M´ARIO A. T. FIGUEIREDO ROBERT D. NOWAK STEPHEN.

Slides:



Advertisements
Similar presentations
Nonnegative Matrix Factorization with Sparseness Constraints S. Race MA591R.
Advertisements

Numerical Solution of Linear Equations
Regularization David Kauchak CS 451 – Fall 2013.
Engineering Optimization
P. Venkataraman Mechanical Engineering P. Venkataraman Rochester Institute of Technology DETC2013 – 12269: Continuous Solution for Boundary Value Problems.
Siddharth Choudhary.  Refines a visual reconstruction to produce jointly optimal 3D structure and viewing parameters  ‘bundle’ refers to the bundle.
Optimization of thermal processes2007/2008 Optimization of thermal processes Maciej Marek Czestochowa University of Technology Institute of Thermal Machinery.
Exact or stable image\signal reconstruction from incomplete information Project guide: Dr. Pradeep Sen UNM (Abq) Submitted by: Nitesh Agarwal IIT Roorkee.
Wangmeng Zuo, Deyu Meng, Lei Zhang, Xiangchu Feng, David Zhang
Inexact SQP Methods for Equality Constrained Optimization Frank Edward Curtis Department of IE/MS, Northwestern University with Richard Byrd and Jorge.
Analysis of the Increase and Decrease Algorithms for Congestion Avoidance in Computer Networks Dah-Ming Chiu and Raj Jain Presented by Yao Zhao.
“ Analysis of the Increase and Decrease Algorithms for Congestion Avoidance in Computer Networks ”
Gizem ALAGÖZ. Simulation optimization has received considerable attention from both simulation researchers and practitioners. Both continuous and discrete.
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Numerical Optimization
Function Optimization Newton’s Method. Conjugate Gradients
Radial Basis Functions
Nonlinear Optimization for Optimal Control
Motion Analysis (contd.) Slides are from RPI Registration Class.
Methods For Nonlinear Least-Square Problems
CSci 6971: Image Registration Lecture 4: First Examples January 23, 2004 Prof. Chuck Stewart, RPI Dr. Luis Ibanez, Kitware Prof. Chuck Stewart, RPI Dr.
Linear Regression  Using a linear function to interpolate the training set  The most popular criterion: Least squares approach  Given the training set:
The Widrow-Hoff Algorithm (Primal Form) Repeat: Until convergence criterion satisfied return: Given a training set and learning rate Initial:  Minimize.
The Perceptron Algorithm (Dual Form) Given a linearly separable training setand Repeat: until no mistakes made within the for loop return:
Reformulated - SVR as a Constrained Minimization Problem subject to n+1+2m variables and 2m constrains minimization problem Enlarge the problem size and.
Active Set Support Vector Regression
Support Vector Regression David R. Musicant and O.L. Mangasarian International Symposium on Mathematical Programming Thursday, August 10, 2000
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Application of Digital Signal Processing in Computed tomography (CT)

Computational Optimization
Collaborative Filtering Matrix Factorization Approach
Online Dictionary Learning for Sparse Coding International Conference on Machine Learning, 2009 Julien Mairal, Francis Bach, Jean Ponce and Guillermo Sapiro.
Algorithm Taxonomy Thus far we have focused on:
AN ITERATIVE METHOD FOR MODEL PARAMETER IDENTIFICATION 4. DIFFERENTIAL EQUATION MODELS E.Dimitrova, Chr. Boyadjiev E.Dimitrova, Chr. Boyadjiev BULGARIAN.
Primal-Dual Meets Local Search: Approximating MST’s with Non-uniform Degree Bounds Author: Jochen Könemann R. Ravi From CMU CS 3150 Presentation by Dan.
Image Restoration using Iterative Wiener Filter --- ECE533 Project Report Jing Liu, Yan Wu.
1 RECENT DEVELOPMENTS IN MULTILAYER PERCEPTRON NEURAL NETWORKS Walter H. Delashmit Lockheed Martin Missiles and Fire Control Dallas, TX 75265
Eran Treister and Irad Yavneh Computer Science, Technion (with thanks to Michael Elad)
Stochastic Subgradient Approach for Solving Linear Support Vector Machines Jan Rupnik Jozef Stefan Institute.
Unit-V DSP APPLICATIONS. UNIT V -SYLLABUS DSP APPLICATIONS Multirate signal processing: Decimation Interpolation Sampling rate conversion by a rational.
Computer Animation Rick Parent Computer Animation Algorithms and Techniques Optimization & Constraints Add mention of global techiques Add mention of calculus.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
Regularization and Feature Selection in Least-Squares Temporal Difference Learning J. Zico Kolter and Andrew Y. Ng Computer Science Department Stanford.
Vaida Bartkutė, Leonidas Sakalauskas
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Exact Differentiable Exterior Penalty for Linear Programming Olvi Mangasarian UW Madison & UCSD La Jolla Edward Wild UW Madison December 20, 2015 TexPoint.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Inexact SQP methods for equality constrained optimization Frank Edward Curtis Department of IE/MS, Northwestern University with Richard Byrd and Jorge.
Chapter 2-OPTIMIZATION
Variations on Backpropagation.
Mathematical Analysis of MaxEnt for Mixed Pixel Decomposition
METHOD OF STEEPEST DESCENT ELE Adaptive Signal Processing1 Week 5.
Machine learning optimization Usman Roshan. Machine learning Two components: – Modeling – Optimization Modeling – Generative: we assume a probabilistic.
Optimization in Engineering Design 1 Introduction to Non-Linear Optimization.
Li-Wei Kang and Chun-Shien Lu Institute of Information Science, Academia Sinica Taipei, Taiwan, ROC {lwkang, April IEEE.
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
Massive Support Vector Regression (via Row and Column Chunking) David R. Musicant and O.L. Mangasarian NIPS 99 Workshop on Learning With Support Vectors.
Regularized Least-Squares and Convex Optimization.
Bounded Nonlinear Optimization to Fit a Model of Acoustic Foams
Compressive Coded Aperture Video Reconstruction
Jeremy Watt and Aggelos Katsaggelos Northwestern University
Multiplicative updates for L1-regularized regression
Probabilistic Models for Linear Regression
NESTA: A Fast and Accurate First-Order Method for Sparse Recovery
Collaborative Filtering Matrix Factorization Approach
6.5 Taylor Series Linearization
Multiple features Linear Regression with multiple variables
Multiple features Linear Regression with multiple variables
Survey on Coverage Problems in Wireless Sensor Networks - 2
Presentation transcript:

GRADIENT PROJECTION FOR SPARSE RECONSTRUCTION: APPLICATION TO COMPRESSED SENSING AND OTHER INVERSE PROBLEMS M´ARIO A. T. FIGUEIREDO ROBERT D. NOWAK STEPHEN J. WRIGHT

BACKGROUND

PREVIOUS ALGORITHMS Interior-point method SparseLab: a Matlab software package designed to find sparse solutions to systems of linear equations L1_ls: a Matlab implementation of the interior-point method for L1-regularized least squares L1-MAGIC: a collection of MATLAB routines for solving the convex optimization programs central to compressive sampling

GRADIENT PROJECTION FOR SPARSE RECONSTRUCTION Formulation

GRADIENT DESCENT Gradient descent is a first-order optimization algorithm. To find a local minimum of a function using gradient descent, one takes steps proportional to the negative of the gradient (or of the approximate gradient) of the function at the current point.

BASIC GRADIENT PROJECTION

GPSR-BASIC

GPSR-BB An approach due to Barzilai and Borwein (BB) that F may not decrease at every iteration. Standard non-monotone method: Eliminate the backtracking line search step Monotone method:

CONVERGENCE Theorem 1: The sequence of iterates {z (k) } generated by the either the GPSR-Basic of GPSR-BB algorithms either terminates at a solution or else converges to a solution of function below at an R-linear rate. T. Serafini, G. Zanghirati, L. Zanni. “Gradient projection methods for large quadratic programs and applications in training support vector machines,” Optimization Methods and Software, vol. 20, pp. 353–378, 2004.

TERMINATION Several termination criterions are presented while these options all perform well on all data sets. The one author used in this paper is motivated by perturbation results for linear complementarity problem (LCP).

DEBIASING The function above is a bias solution for least-square problem. So we could fix the zero components, and use standard least square to get a debiasing solution. It is also worth pointing out that debiasing is not always desirable. Shrinking the selected coefficients can mitigate unusually large noise deviations, a desirable effect that may be undone by debiasing.

WARM STARTING AND CONTINUATION After solving the problem with a given τ, we could use the solution to initialize GPSR for a nearby value of τ. It has been noted recently that the speed of GPSR may degrade considerably for smaller values of the regularization parameter τ. However, if we use GPSR for a larger value of τ, then decrease τ in steps toward its desired value. To benefit from a warm start, IP methods require the initial point to be not only close to the solution but also sufficiently interior to the feasible set and close to a “central path,” which is difficult to satisfy in practice.

EXPERIMENTS

CONCLUSIONS Significantly faster than the state-of-the-art algorithms in experimental comparisons Poor performance when the regularization parameter τ is small, while continuation heuristic could be used to recover efficient practical performance While it is not obvious WHY GPSR perform well