QUASI MAXIMUM LIKELIHOOD BLIND DECONVOLUTION QUASI MAXIMUM LIKELIHOOD BLIND DECONVOLUTION Alexander Bronstein.

Slides:



Advertisements
Similar presentations
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The Linear Prediction Model The Autocorrelation Method Levinson and Durbin.
Advertisements

1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
2 – In previous chapters: – We could design an optimal classifier if we knew the prior probabilities P(wi) and the class- conditional probabilities P(x|wi)
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
Chap 8: Estimation of parameters & Fitting of Probability Distributions Section 6.1: INTRODUCTION Unknown parameter(s) values must be estimated before.
2008 SIAM Conference on Imaging Science July 7, 2008 Jason A. Palmer
CWIT Robust Entropy Rate for Uncertain Sources: Applications to Communication and Control Systems Charalambos D. Charalambous Dept. of Electrical.
Maximum likelihood separation of spatially autocorrelated images using a Markov model Shahram Hosseini 1, Rima Guidara 1, Yannick Deville 1 and Christian.
ECE 472/572 - Digital Image Processing Lecture 8 - Image Restoration – Linear, Position-Invariant Degradations 10/10/11.
Visual Recognition Tutorial
Pattern Recognition and Machine Learning
Numerical Optimization
Maximum likelihood (ML) and likelihood ratio (LR) test
AGC DSP AGC DSP Professor A G Constantinides© Estimation Theory We seek to determine from a set of data, a set of parameters such that their values would.
SYSTEMS Identification
Ambiguity in Radar and Sonar Paper by M. Joao D. Rendas and Jose M. F. Moura Information theory project presented by VLAD MIHAI CHIRIAC.
Newton Method for the ICA Mixture Model
Parametric Inference.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Independent Component Analysis (ICA) and Factor Analysis (FA)
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Visual Recognition Tutorial
Maximum likelihood (ML)
Advanced Image Processing Image Relaxation – Restoration and Feature Extraction 02/02/10.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
ELE 488 F06 ELE 488 Fall 2006 Image Processing and Transmission ( ) Wiener Filtering Derivation Comments Re-sampling and Re-sizing 1D  2D 10/5/06.
Introduction to Adaptive Digital Filters Algorithms
Chapter 15 Modeling of Data. Statistics of Data Mean (or average): Variance: Median: a value x j such that half of the data are bigger than it, and half.
Speech Recognition Pattern Classification. 22 September 2015Veton Këpuska2 Pattern Classification  Introduction  Parametric classifiers  Semi-parametric.
Independent Component Analysis Zhen Wei, Li Jin, Yuxue Jin Department of Statistics Stanford University An Introduction.
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Semi-Blind (SB) Multiple-Input Multiple-Output (MIMO) Channel Estimation Aditya K. Jagannatham DSP MIMO Group, UCSD ArrayComm Presentation.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Definitions Random Signal Analysis (Review) Discrete Random Signals Random.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
Clustering and Testing in High- Dimensional Data M. Radavičius, G. Jakimauskas, J. Sušinskas (Institute of Mathematics and Informatics, Vilnius, Lithuania)
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
: Chapter 3: Maximum-Likelihood and Baysian Parameter Estimation 1 Montri Karnjanadecha ac.th/~montri.
Consistency An estimator is a consistent estimator of θ, if , i.e., if
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 07: BAYESIAN ESTIMATION (Cont.) Objectives:
High-dimensional Error Analysis of Regularized M-Estimators Ehsan AbbasiChristos ThrampoulidisBabak Hassibi Allerton Conference Wednesday September 30,
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
M.Sc. in Economics Econometrics Module I Topic 4: Maximum Likelihood Estimation Carol Newman.
Diversity Loss in General Estimation of Distribution Algorithms J. L. Shapiro PPSN (Parallel Problem Solving From Nature) ’06 BISCuit 2 nd EDA Seminar.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Week 31 The Likelihood Function - Introduction Recall: a statistical model for some data is a set of distributions, one of which corresponds to the true.
Yi Jiang MS Thesis 1 Yi Jiang Dept. Of Electrical and Computer Engineering University of Florida, Gainesville, FL 32611, USA Array Signal Processing in.
EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
Numerical Analysis – Data Fitting Hanyang University Jong-Il Park.
State-Space Recursive Least Squares with Adaptive Memory College of Electrical & Mechanical Engineering National University of Sciences & Technology (NUST)
Using Neumann Series to Solve Inverse Problems in Imaging Christopher Kumar Anand.
Presentation : “ Maximum Likelihood Estimation” Presented By : Jesu Kiran Spurgen Date :
12. Principles of Parameter Estimation
Probability Theory and Parameter Estimation I
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
Tirza Routtenberg Dept. of ECE, Ben-Gurion University of the Negev
Ch3: Model Building through Regression
Department of Civil and Environmental Engineering
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Computing and Statistical Data Analysis / Stat 7
Learning Theory Reza Shadmehr
Mathematical Foundations of BME
12. Principles of Parameter Estimation
Presentation transcript:

QUASI MAXIMUM LIKELIHOOD BLIND DECONVOLUTION QUASI MAXIMUM LIKELIHOOD BLIND DECONVOLUTION Alexander Bronstein

BIBLIOGRAPHY 2 2 A. Bronstein, M. Bronstein, and M. Zibulevsky, "Quasi maximum likelihood blind deconvolution: super- ans sub-Gaussianity vs. asymptotic stability", submitted to IEEE Trans. Sig. Proc. A. Bronstein, M. Bronstein, M. Zibulevsky, and Y. Y. Zeevi, "Quasi maximum likelihood blind deconvolution: asymptotic performance analysis", submitted to IEEE Trans. Information Theory. A. Bronstein, M. Bronstein, and M. Zibulevsky, "Relative optimization for blind deconvolution", submitted to IEEE Trans. Sig. Proc. A. Bronstein, M. Bronstein, M. Zibulevsky, and Y. Y. Zeevi, "Quasi maximum likelihood blind deconvolution of images acquired through scattering media", Submitted to ISBI04. A. Bronstein, M. Bronstein, M. Zibulevsky, and Y. Y. Zeevi, "Quasi maximum likelihood blind deconvolution of images using optimal sparse representations", CCIT Report No. 455 (EE No. 1399), Dept. of Electrical Engineering, Technion, Israel, December A. Bronstein, M. Bronstein, and M. Zibulevsky, "Blind deconvolution with relative Newton method", CCIT Report No. 444 (EE No. 1385), Dept. of Electrical Engineering, Technion, Israel, October 2003.

AGENDA 3 3 Introduction QML blind deconvolution Asymptotic analysis Relative Newton Generalizations Problem formulation Applications

BLIND DECONVOLUTION PROBLEM 4 4 source signal convolution kernel observed signal sensor noise signal WH CONVOLUTION MODELDECONVOLUTION restoration kernel source estimate arbitrary scaling factor arbitrary delay

APPLICATIONS 5 5 Acoustics, speech processing  DEREVERBERATION Optics, image processing, biomedical imaging  DEBLURRING Communications  CHANNEL EQUALIZATION Control  SYSTEM IDENTIFICATION Statistics, finances  ARMA ESTIMATION

AGENDA 6 6 Introduction QML blind deconvolution Asymptotic analysis Relative Newton Generalizations ML vs. QML The choice of φ(s) Equivariance Gradient and Hessian

ML BLIND DECONVOLUTION is i.i.d. with probability density function 2 has no zeros on the unit circle, i.e. ASSUMPTIONS MAXIMUM-LIKELIHOOD BLIND DECONVOLUTION: 3 No noise (precisely: no noise model) 4 is zero-mean

QUASI ML BLIND DECONVOLUTION 8 8 The true source PDF in usually unknown Many times is non-log-concave and not well-suited for optimization Substitute with some model function PROBLEMS OF MAXIMUM LIKELIHOOD QUASI MAXIMUM LIKELIHOOD BLIND DECONVOLUTION

THE CHOICE OF  (s) 9 9 SUPER-GAUSSIANSUB-GAUSSIAN

EQUIVARIANCE 10 QML estimator of given the observation Theorem: The QML estimator is equivariant, i.e., for every invertible kernel, it holds where stands for the impulse response of the inverse of. ANALYSIS OF

GRADIENT & HESSIAN OF 11 GRADIENT where is the mirror operator. HESSIAN

AGENDA 12 Introduction QML blind deconvolution Asymptotic analysis Relative Newton Generalizations Asymptotic Hessian structure Asymptotic error covariance Cramér-Rao bounds Superefficiency Examples

ASYMPTOTIC HESSIAN AT THE SOLUTION POINT 13 For a sufficiently large sample size, the Hessian becomes At the solution point, where

ASYMPTOTIC ERROR COVARIANCE 14 Estimation kernel from the data Exact restoration kernel The scaling factor has to obey

ASYMPTOTIC ERROR COVARIANCE 15 Estimation error From second-order Taylor expansion, equivariance

ASYMPTOTIC ERROR COVARIANCE 16 Asymptotically ( ), Separable structure:

ASYMPTOTIC ERROR COVARIANCE 17 The estimation error covariance matrix asymptotically separates to

ASYMPTOTIC ERROR COVARIANCE 18 Asymptotic gradient covariance matrices where

ASYMPTOTIC ERROR COVARIANCE 19 Asymptotic signal-to-interference ratio (SIR) estimate: Asymptotic estimation error covariance:

CRAMER-RAO LOWER BOUNDS 20 True ML estimator: The distribution-dependent parameters simplify to Asymptotic error covariance simplifies to where

CRAMER-RAO LOWER BOUNDS 21 Asymptotic SIR estimate simplifies to

SUPEREFFICIENCY 22 Let the source be sparse, i.e., Let be the smoothed absolute value with smoothing parameter In the limit

SUPEREFFICIENCY 23 Similar results are obtained for uniformly-distributed source with Can be extended for sources with PDF vanishing outside some interval.

ASYMPTOTIC STABILITY 24 The QML estimator is said to be asymptotically stable if is a local minimizer of in the limit. Theorem: The QML estimator is asymptotically stable if the following conditions hold: and is asymptotically unstable if one of the following conditions hold:

EXAMPLE 25 Generalized Laplace distribution

STABILITY OF THE SUPER-GAUSSIAN ESTIMATOR 26 SUPER-GAUSSIAN SUB-GAUSSIAN

STABILITY OF THE SUB-GAUSSIAN ESTIMATOR 27 SUPER-GAUSSIAN SUB-GAUSSIAN

PERFORMANCE OF THE SUPER-GAUSSIAN ESTIMATOR 28 SUPER-GAUSSIAN

PERFORMANCE OF THE SUB-GAUSSIAN ESTIMATOR 29 SUB-GAUSSIAN

AGENDA 30 Introduction QML blind deconvolution Asymptotic analysis Relative Newton Generalizations Relative optimization Relative Newton Fast Relative Newton

RELATIVE OPTIMIZATION (RO) 31 0 Start with and 1 For until convergence 4 Update source estimate 2 Start with 3 Find such that 5 End For Restoration kernel estimate: Source estimate:

RELATIVE OPTIMIZATION (RO) 32 Observation: The k-th step of the relative optimization algorithm depends only on Proposition: The sequence of target function values produced by the relative optimization algorithm is monotonically decreasing, i.e.,

RELATIVE NEWTON 33 Relative Newton = use one Newton step in the RO algorithm Near the solution point Newton system separates to

FAST RELATIVE NEWTON 34 Fast relative Newton = use one Newton step with approximate Hessian in the RO algorithm + regularized approximate Newton system solution. Approximate Hessian evaluation = order of gradient evaluation

AGENDA 35 Introduction QML blind deconvolution Asymptotic analysis Relative Newton Generalizations

GENERALIZATIONS 36 IIR KERNELS BLOCK PROCESSING  ONLINE DECONVOLUTION MULTI-CHANNEL DECONVOLUTION  BSS+BD DECONVOLUTION OF IMAGES + USE OF SPARSE REPRESENTATIONS

GENERALIZATIONS: IIR KERNELS IIR FIR

GENERALIZATIONS: ONLINE PROCESSING Fast Relative Newton (block)

GENERALIZATIONS: DECONVOLUTION OF IMAGES SOURCEOBSERVATIONRESTORATION

GENERALIZATIONS: DECONVOLUTION OF IMAGES Fast relative Newton Newton Fast relative Newton