Fast and incoherent dictionary learning algorithms with application to fMRI Authors: Vahid Abolghasemi Saideh Ferdowsi Saeid Sanei. Journal of Signal Processing.

Slides:



Advertisements
Similar presentations
Active Shape Models Suppose we have a statistical shape model –Trained from sets of examples How do we use it to interpret new images? Use an “Active Shape.
Advertisements

Structured Sparse Principal Component Analysis Reading Group Presenter: Peng Zhang Cognitive Radio Institute Friday, October 01, 2010 Authors: Rodolphe.
Learning Measurement Matrices for Redundant Dictionaries Richard Baraniuk Rice University Chinmay Hegde MIT Aswin Sankaranarayanan CMU.
Fast Bayesian Matching Pursuit Presenter: Changchun Zhang ECE / CMR Tennessee Technological University November 12, 2010 Reading Group (Authors: Philip.
T HE POWER OF C ONVEX R ELAXATION : N EAR - OPTIMAL MATRIX COMPLETION E MMANUEL J. C ANDES AND T ERENCE T AO M ARCH, 2009 Presenter: Shujie Hou February,
Patch-based Image Deconvolution via Joint Modeling of Sparse Priors Chao Jia and Brian L. Evans The University of Texas at Austin 12 Sep
K-SVD Dictionary-Learning for Analysis Sparse Models
Learning sparse representations to restore, classify, and sense images and videos Guillermo Sapiro University of Minnesota Supported by NSF, NGA, NIH,
2008 SIAM Conference on Imaging Science July 7, 2008 Jason A. Palmer
Extensions of wavelets
* * Joint work with Michal Aharon Freddy Bruckstein Michael Elad
1 Micha Feigin, Danny Feldman, Nir Sochen
More MR Fingerprinting
An Introduction to Sparse Coding, Sparse Sensing, and Optimization Speaker: Wei-Lun Chao Date: Nov. 23, 2011 DISP Lab, Graduate Institute of Communication.
1 Applications on Signal Recovering Miguel Argáez Carlos A. Quintero Computational Science Program El Paso, Texas, USA April 16, 2009.
Bayesian Robust Principal Component Analysis Presenter: Raghu Ranganathan ECE / CMR Tennessee Technological University January 21, 2011 Reading Group (Xinghao.
Entropy-constrained overcomplete-based coding of natural images André F. de Araujo, Maryam Daneshi, Ryan Peng Stanford University.
1cs542g-term Notes  In assignment 1, problem 2: smoothness = number of times differentiable.
1cs542g-term Notes  Assignment 1 due tonight ( me by tomorrow morning)
Dictionary-Learning for the Analysis Sparse Model Michael Elad The Computer Science Department The Technion – Israel Institute of technology Haifa 32000,
Sparse and Overcomplete Data Representation
Some useful linear algebra. Linearly independent vectors span(V): span of vector space V is all linear combinations of vectors v i, i.e.
Image Denoising via Learned Dictionaries and Sparse Representations
Optimized Projection Directions for Compressed Sensing Michael Elad The Computer Science Department The Technion – Israel Institute of technology Haifa.
Real-time Combined 2D+3D Active Appearance Models Jing Xiao, Simon Baker,Iain Matthew, and Takeo Kanade CVPR 2004 Presented by Pat Chan 23/11/2004.
Image Denoising with K-SVD Priyam Chatterjee EE 264 – Image Processing & Reconstruction Instructor : Prof. Peyman Milanfar Spring 2007.
Independent Component Analysis (ICA) and Factor Analysis (FA)
Recent Trends in Signal Representations and Their Role in Image Processing Michael Elad The CS Department The Technion – Israel Institute of technology.
The Quest for a Dictionary. We Need a Dictionary  The Sparse-land model assumes that our signal x can be described as emerging from the PDF:  Clearly,
MATH 685/ CSI 700/ OR 682 Lecture Notes Lecture 6. Eigenvalue problems.
Dominant Eigenvalues & The Power Method
Normalised Least Mean-Square Adaptive Filtering

AMSC 6631 Sparse Solutions of Linear Systems of Equations and Sparse Modeling of Signals and Images: Midyear Report Alfredo Nava-Tudela John J. Benedetto,
Chapter 15 Modeling of Data. Statistics of Data Mean (or average): Variance: Median: a value x j such that half of the data are bigger than it, and half.
1 Iterative Solution Methods Starts with an initial approximation for the solution vector (x 0 ) At each iteration updates the x vector by using the sytem.
Cs: compressed sensing
 Karthik Gurumoorthy  Ajit Rajwade  Arunava Banerjee  Anand Rangarajan Department of CISE University of Florida 1.
Orthogonalization via Deflation By Achiya Dax Hydrological Service Jerusalem, Israel
Learning Theory Reza Shadmehr LMS with Newton-Raphson, weighted least squares, choice of loss function.
Learning to Sense Sparse Signals: Simultaneous Sensing Matrix and Sparsifying Dictionary Optimization Julio Martin Duarte-Carvajalino, and Guillermo Sapiro.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
Fields of Experts: A Framework for Learning Image Priors (Mon) Young Ki Baik, Computer Vision Lab.
Sparse Signals Reconstruction Via Adaptive Iterative Greedy Algorithm Ahmed Aziz, Ahmed Salim, Walid Osamy Presenter : 張庭豪 International Journal of Computer.
Journal Club Journal of Chemometrics May 2010 August 23, 2010.
Image Decomposition, Inpainting, and Impulse Noise Removal by Sparse & Redundant Representations Michael Elad The Computer Science Department The Technion.
Spectrum Sensing In Cognitive Radio Networks
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
CS Statistical Machine learning Lecture 12 Yuan (Alan) Qi Purdue CS Oct
Ultra-high dimensional feature selection Yun Li
Single Image Interpolation via Adaptive Non-Local Sparsity-Based Modeling The research leading to these results has received funding from the European.
Jianchao Yang, John Wright, Thomas Huang, Yi Ma CVPR 2008 Image Super-Resolution as Sparse Representation of Raw Image Patches.
From Sparse Solutions of Systems of Equations to Sparse Modeling of Signals and Images Alfred M. Bruckstein (Technion), David L. Donoho (Stanford), Michael.
Computational Physics (Lecture 7) PHY4061. Eigen Value Problems.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Iterative Solution Methods
Jinbo Bi Joint work with Jiangwen Sun, Jin Lu, and Tingyang Xu
Author: Vikas Sindhwani and Amol Ghoting Presenter: Jinze Li
Random walk initialization for training very deep feedforward networks
Presented by Nagesh Adluru
Parallelization of Sparse Coding & Dictionary Learning
Learning Theory Reza Shadmehr
Improving K-SVD Denoising by Post-Processing its Method-Noise
* * Joint work with Michal Aharon Freddy Bruckstein Michael Elad
Sparse and Redundant Representations and Their Applications in
NON-NEGATIVE COMPONENT PARTS OF SOUND FOR CLASSIFICATION Yong-Choon Cho, Seungjin Choi, Sung-Yang Bang Wen-Yi Chu Department of Computer Science &
Learned Convolutional Sparse Coding
Outline Sparse Reconstruction RIP Condition
Progress Report Alvaro Velasquez.
Presentation transcript:

Fast and incoherent dictionary learning algorithms with application to fMRI Authors: Vahid Abolghasemi Saideh Ferdowsi Saeid Sanei. Journal of Signal Processing Systems /04/11

Content 1.K-SVD 2.Incoherent K-SVD (IK-SVD) 3.Fast incoherent dictionary learning (FIDL) 4.Results

K-SVD Assume that the signal can be represented as a linear combination of a few atoms in dictionary such that. The DL problem can be expressed as:

K-SVD a)Sparse coding (Keep D fixed, update X) assume that τ is known and apply OMP to solve: b)Updating D

K-SVD In order to minimize (3), apply SVD to and simultaneously update and using the strongest eigenvector and eigenvalue of

IK-SVD High incoherence between the dictionary atoms is desired in almost all dictionary learning methods. This guarantees that the atoms are as discriminative as possible. A remedy—a suitable tool for evaluating the coherence between the atoms is Gram matrix :

IK-SVD Matrix G is K×K and symmetric with unit diagonal elements (note that D is column- normalized). The absolute values of off-diagonal elements of G represent the degree of coherence between any pair of atoms in D and therefore are desired to be very small.

IK-SVD In order to minimize the above problem, we first take the gradient of which is computed :

IK-SVD where γ = 4ξ > 0 is the step size controlling the convergence behavior of the algorithm, and k is the iteration counter of the incoherence constraint stage.

IK-SVD

FIDL Because of computationally expensive for learning large dictionaries. In order to design to be fast and at the same time exploits the incoherence of atoms.

FIDL Use - norm of the entire matrix X defined as instead of forcing individual vectors to be sparse. This allows us to update the coefficients, simultaneously rather than column-by-column. The incoherence constraint on D is also added to the cost function.

FIDL a)Coefficient update (sparse coding) split into the sum of a smooth and a nonsmooth sub-cost function, represented by P and Q. Gradient descent step: Proximal step:

FIDL The proximal function here is defined using soft- thresholding (Shrink{.}), which ultimately leads to: b) Dictionary update

FIDL

Results-Experiment 1 The nonzero entries of the 20×1,000 sparse matrix X were generated randomly (from Gaussian distribution). D was selected as a random overcomplete full-rank matrix of size 15×20 with all columns normalized to one.

Results-Experiment 1

Results-Experiment 2 To learn an over-complete dictionary of size 64 × 256 over 14,000 noisy image patches of size 8 × 8 extracted from Barbara image.

Results-Experiment 3 The aim was to investigate the robustness of the proposed methods against variations in the input noise. We considered noisy model Y = DX + V, where V was Gaussian noise with zero mean. All matrices were drawn randomly (from Gaussian distribution) with n = 15, K = 20, and N = 1,000. The number of nonzeros at each column of X was set to five.

Results-Experiment 3

Results-Experiment 4 To evaluate the computational cost of the proposed methods and comparing these methods with other well-known algorithms. The parameters for the algorithms were similar to the first experiment. However, we increased the dictionary size from 5×10 to 500×1,000 for a fixed level of sparsity τ = 2.

Results-Experiment 4

Results-Synthetic fMRI data The simulations started by forming X of size 5×3,600 using five vectorized source images of size 60 × 60. The mixtures were generated by multiplying column-normalized D of size 100 × 5 by X.

Results-Synthetic fMRI data

Results-Real fMRI data A real auditory fMRI dataset was considered for this experiment. Chose K = 35 sources

Results-Real fMRI data FIDL IK-SVD

Results-Real fMRI data K-SVD FastICA

Results-Real fMRI data Lp-norm-based method SPM

Thank you!