Adv DSP Spring-2015 Lecture#9 Optimum Filters (Ch:7) Wiener Filters.

Slides:



Advertisements
Similar presentations
State Space Models. Let { x t :t T} and { y t :t T} denote two vector valued time series that satisfy the system of equations: y t = A t x t + v t (The.
Advertisements

ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The Linear Prediction Model The Autocorrelation Method Levinson and Durbin.
Lecture 7 Linear time invariant systems
AGC DSP AGC DSP Professor A G Constantinides©1 Modern Spectral Estimation Modern Spectral Estimation is based on a priori assumptions on the manner, the.
Adaptive Filters S.B.Rabet In the Name of GOD Class Presentation For The Course : Custom Implementation of DSP Systems University of Tehran 2010 Pages.
Ch11 Curve Fitting Dr. Deshi Ye
OPTIMUM FILTERING.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The FIR Adaptive Filter The LMS Adaptive Filter Stability and Convergence.
ELE Adaptive Signal Processing
AGC DSP AGC DSP Professor A G Constantinides©1 A Prediction Problem Problem: Given a sample set of a stationary processes to predict the value of the process.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Newton’s Method Application to LMS Recursive Least Squares Exponentially-Weighted.
Goals of Adaptive Signal Processing Design algorithms that learn from training data Algorithms must have good properties: attain good solutions, simple.
1 Speech Enhancement Wiener Filtering: A linear estimation of clean signal from the noisy signal Using MMSE criterion.
Adaptive FIR Filter Algorithms D.K. Wise ECEN4002/5002 DSP Laboratory Spring 2003.
EE513 Audio Signals and Systems Wiener Inverse Filter Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Adaptive Signal Processing
Linear Prediction Problem: Forward Prediction Backward Prediction
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Adaptive Noise Cancellation ANC W/O External Reference Adaptive Line Enhancement.
RLSELE Adaptive Signal Processing 1 Recursive Least-Squares (RLS) Adaptive Filters.
Computer Vision - Restoration Hanyang University Jong-Il Park.
By Asst.Prof.Dr.Thamer M.Jamel Department of Electrical Engineering University of Technology Baghdad – Iraq.
Week 2ELE Adaptive Signal Processing 1 STOCHASTIC PROCESSES AND MODELS.
Random Processes ECE460 Spring, Power Spectral Density Generalities : Example: 2.
1 Part 5 Response of Linear Systems 6.Linear Filtering of a Random Signals 7.Power Spectrum Analysis 8.Linear Estimation and Prediction Filters 9.Mean-Square.
Real time DSP Professors: Eng. Julian Bruno Eng. Mariano Llamedo Soria.
Mathematical Preliminaries. 37 Matrix Theory Vectors nth element of vector u : u(n) Matrix mth row and nth column of A : a(m,n) column vector.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
1 Linear Prediction. 2 Linear Prediction (Introduction) : The object of linear prediction is to estimate the output sequence from a linear combination.
1 Linear Prediction. Outline Windowing LPC Introduction to Vocoders Excitation modeling  Pitch Detection.
CHAPTER 4 Adaptive Tapped-delay-line Filters Using the Least Squares Adaptive Filtering.
Intro. ANN & Fuzzy Systems Lecture 26 Modeling (1): Time Series Prediction.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Definitions Random Signal Analysis (Review) Discrete Random Signals Random.
Unit-V DSP APPLICATIONS. UNIT V -SYLLABUS DSP APPLICATIONS Multirate signal processing: Decimation Interpolation Sampling rate conversion by a rational.
2. Stationary Processes and Models
EE513 Audio Signals and Systems
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Derivation Computational Simplifications Stability Lattice Structures.
Robotics Research Laboratory 1 Chapter 7 Multivariable and Optimal Control.
Chapter 11 Filter Design 11.1 Introduction 11.2 Lowpass Filters
1 EE571 PART 4 Classification of Random Processes Huseyin Bilgekul Eeng571 Probability and astochastic Processes Department of Electrical and Electronic.
Lecture#10 Spectrum Estimation
Dept. E.E./ESAT-STADIUS, KU Leuven
Professors: Eng. Diego Barral Eng. Mariano Llamedo Soria Julian Bruno
Lecture 10b Adaptive Filters. 2 Learning Objectives  Introduction to adaptive filtering.  LMS update algorithm.  Implementation of an adaptive filter.
Overview of Adaptive Filters Quote of the Day When you look at yourself from a universal standpoint, something inside always reminds or informs you that.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Normal Equations The Orthogonality Principle Solution of the Normal Equations.
Autoregressive (AR) Spectral Estimation
Discrete-time Random Signals
Lecture 12: Parametric Signal Modeling XILIANG LUO 2014/11 1.
Review and Summary Box-Jenkins models Stationary Time series AR(p), MA(q), ARMA(p,q)
Impulse Response Measurement and Equalization Digital Signal Processing LPP Erasmus Program Aveiro 2012 Digital Signal Processing LPP Erasmus Program Aveiro.
Geology 6600/7600 Signal Analysis Last time: Linear Systems Uncorrelated additive noise in the output signal, y = v + n, of a SISO system can be estimated.
Geology 6600/7600 Signal Analysis 23 Oct 2015
CS654: Digital Image Analysis Lecture 22: Image Restoration.
Linear Prediction.
Geology 6600/7600 Signal Analysis 26 Oct 2015 © A.R. Lowry 2015 Last time: Wiener Filtering Digital Wiener Filtering seeks to design a filter h for a linear.
DSP-CIS Part-III : Optimal & Adaptive Filters Chapter-9 : Kalman Filters Marc Moonen Dept. E.E./ESAT-STADIUS, KU Leuven
UNIT-IV. Introduction Speech signal is generated from a system. Generation is via excitation of system. Speech travels through various media. Nature of.
Adv DSP Spring-2015 Lecture#11 Spectrum Estimation Parametric Methods.
ELG5377 Adaptive Signal Processing Lecture 13: Method of Least Squares.
Geology 6600/7600 Signal Analysis Last time: Linear Systems The Frequency Response or Transfer Function of a linear SISO system can be estimated as (Note.
Adaptive Filters Common filter design methods assume that the characteristics of the signal remain constant in time. However, when the signal characteristics.
Pipelined Adaptive Filters
Linear Prediction.
Course: Autonomous Machine Learning
Modern Spectral Estimation
لجنة الهندسة الكهربائية
Image and Video Processing
Linear Prediction.
Wiener Filtering: A linear estimation of clean signal from the noisy signal Using MMSE criterion.
16. Mean Square Estimation
Presentation transcript:

Adv DSP Spring-2015 Lecture#9 Optimum Filters (Ch:7) Wiener Filters

Introduction  Estimation of one signal from another is one of the most important problems in signal processing  In many applications, the desired signal is not available or observed directly. Speech, Radar, EEG etc  Desired signal may be noisy and highly distorted  In very simple and idealized environments, it may be possible to design classical filters such as LP,HP or BP, to restore the desired signal from the measured data.

Introduction  These classical filters shall rarely be optimal in the sense of producing the “best” estimate of the signal.  Class of filters called OPTIMUM DIGITAL FILTERS  Two important Types Digital Wiener Filter Discrete Kalman Filter

Wiener Filter  In the 1940’s, Norbert Wiener pioneered research in the problem of designing a filter that would produce the optimum estimate of a signal from a noisy measurement or observation.

Wiener Filter  The Wiener filtering problem, is to design a filter to recover a signal d[n] from noisy measurement  Assuming that both d[n] and v[n] are wide-sense stationary random process, Wiener considered the problem of designing the filter that would produces the minimum mean square estimate of d[n] (by using x[n])

Wiener Filter  Thus the error terms are (Mean Square Error)  Problem is to find the filter (filter coefficients, FIR or IIR) that minimizes ξ (minimum mean square error).

Wiener Filter  Depending on how the signals x[n] and d[n] are related to each other, a number of different and important problems may be cast into Wiener filtering framework.  These problems are  Filtering  Smoothing  Prediction  De-convolution

Wiener Filter

The FIR Wiener Filter  Design of an FIR Wiener Filter That will produce the minimum mean-square estimate of a given process d[n] by filtering a set of observations of statistically related process x[n]  It is assumed that x[n] and d[n] are jointly wide-sense stationary with known autocorrelations r x [k] and r d [k], and known cross-correlation r dx [k]

The FIR Wiener Filter  Denoting the unit sample response of the Wiener filter by w[n], and assuming (p-1)st order filter, the system function W(z) is  With x[n] as input to this filter, the output is (using DT convolution)

The FIR Wiener Filter  Wiener filter design problem requires that we find the filter coefficients w[k], that minimize the mean-square error:  Using optimization steps, for k=0,1,…,p-1

Error Taking Partial Derivative for k th value Orthogonality Principle After putting e[n] back

The FIR Wiener Filter  Since x[n] and d[n] are jointly WSS  Set of ‘p’ linear equations in the ‘p’ unknowns w[k] for k=0,1,….,p-1

The FIR Wiener Filter  In matrix form using the fact that autocorrelation sequence is conjugate symmetric r x [k]=r* x [-k] Wiener-Hopf Equations

The FIR Wiener Filter  Wiener-Hopf Equations

The FIR Wiener Filter  The minimum mean square error in the estimate of d[n] is Equal to zero bz of Orthognality

The FIR Wiener Filter  After taking expected values In Vector Notation

The FIR Wiener Filter

Filtering  In the filtering problem, a signal d[n] is to be estimated from a noise (v[n]) corrupted observation x[n]  Assuming that noise has a zero mean and it is uncorrelated with d[n]

Filtering  The cross-correlation between d[n] and x[n] becomes

Filtering  With v[n] and d[n] uncorrelated, it follows  To simplify these equations, specific information about the statistic of the signal and noise are required Example:7.2.1

Linear Prediction  With noise-free observations, linear prediction is concerned with the estimation (prediction) of x[n+1] in terms of linear combination of the current and previous values of x[n]

Linear Prediction  An FIR linear predictor of order ‘p-1’ has the form where w[k] for k=0,1,…,p-1 are the coefficients of the prediction filter.  Linear predictor may be cast into Wiener filtering problem by setting d[n]=x[n+1]

Linear Prediction  Setting up the Wiener-Hopf equations Ex:7.2.2

Linear Prediction in noise  With noise present, a more realistic model for linear prediction is the one in which the signal to be predicted is measured in the presence of noise.

Linear Prediction in noise  Input to Wiener filter is given by  Goal is to design a filter that will estimate x[n+1] in terms of linear combination of ‘p’ previous values of y[n]

Linear Prediction in noise  The Wiener-Hopf equations are  If the noise is uncorrelated with signal x[n], then R y, the autocorrelation matrix for y[n] is

Linear Prediction in noise  The only difference between linear prediction with and without noise is in the autocorrelation matrix for the input signal.  In the case of noise that is uncorrelated with x[n],

Multi-Step Prediction  In one-step linear prediction, x[n+1] is predicted in terms of linear combination of the current and previous values of x[n]  In multi-step prediction, x[n+δ] is predicted in terms of linear combination of the ‘p’ values x[n],x[n-1],…,x[n-p+1]

Multi-Step Prediction  In multi-step prediction  In multi-step prediction, since Positive Integer

Multi-Step Prediction  Wiener-Hopf equations are

Noise Cancellation  The goal of noise canceller is to estimate a signal d[n] from a noise corrupted observation, that is recorded by primary sensor.  Unlike the filtering problem, which requires that the autocorrelation of the noise be known, with noise canceller this information is obtained from a secondary sensor that is placed within the noise field.

Noise Cancellation Primary sensor Secondary sensor

Noise Cancellation  Although the noise measured by secondary sensor, v 2 [n], will be correlated with the noise in the primary sensor v 1 [n], the two will not be same.  Reasons for being not same: Difference in sensor characteristics Difference in the propagation path from noise source to the two sensors.  Since v 1 [n]≠v 2 [n], it is not possible to estimate d[n] by simply subtracting v 2 [n] from x[n]

Noise Cancellation  Noise canceller consists of Wiener filter that is designed to estimate the noise v 1 [n] from the signal received by the secondary sensor  This estimate is then subtracted from the primary signal x[n] to form an estimate of d[n] which is given by

Noise Cancellation  With v 2 [n] as the input to Wiener filter, that is used to estimate the noise v 1 [n], the Wiener-Hopf equations are  R v 2 is the autocorrelation matrix of v 2 [n]  r v 1,v2 is the vector of cross-correlations between desired signal v 1 [n] and Wiener filter input v 2 [n]

Noise Cancellation  The cross-correlation between v 1 [n] and v 2 [n] is  If we assume v 2 [n] is uncorrelated with d[n], then second term is zero, hence

Example:7.2.6