1 Linear Prediction. 2 Linear Prediction (Introduction) : The object of linear prediction is to estimate the output sequence from a linear combination.

Slides:



Advertisements
Similar presentations
Dates for term tests Friday, February 07 Friday, March 07
Advertisements

ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The Linear Prediction Model The Autocorrelation Method Levinson and Durbin.
Liner Predictive Pitch Synchronization Voiced speech detection, analysis and synthesis Jim Bryan Florida Institute of Technology ECE5525 Final Project.
AGC DSP AGC DSP Professor A G Constantinides©1 Modern Spectral Estimation Modern Spectral Estimation is based on a priori assumptions on the manner, the.
Speech Recognition Chapter 3
OPTIMUM FILTERING.
Speaker Recognition Sharat.S.Chikkerur Center for Unified Biometrics and Sensors
A 12-WEEK PROJECT IN Speech Coding and Recognition by Fu-Tien Hsiao and Vedrana Andersen.
Itay Ben-Lulu & Uri Goldfeld Instructor : Dr. Yizhar Lavner Spring /9/2004.
Speech and Audio Processing and Recognition
ELE Adaptive Signal Processing
AGC DSP AGC DSP Professor A G Constantinides©1 A Prediction Problem Problem: Given a sample set of a stationary processes to predict the value of the process.
Speech Coding Nicola Orio Dipartimento di Ingegneria dell’Informazione IV Scuola estiva AISV, 8-12 settembre 2008.
Overview of Adaptive Multi-Rate Narrow Band (AMR-NB) Speech Codec
Pole Zero Speech Models Speech is nonstationary. It can approximately be considered stationary over short intervals (20-40 ms). Over thisinterval the source.
Voice Transformation Project by: Asaf Rubin Michael Katz Under the guidance of: Dr. Izhar Levner.
Adaptive FIR Filter Algorithms D.K. Wise ECEN4002/5002 DSP Laboratory Spring 2003.
EE513 Audio Signals and Systems Wiener Inverse Filter Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Linear Prediction Problem: Forward Prediction Backward Prediction
STAT 497 LECTURE NOTES 2.
Linear Prediction Coding (LPC)
1 CS 551/651: Structure of Spoken Language Lecture 8: Mathematical Descriptions of the Speech Signal John-Paul Hosom Fall 2008.
Digital Systems: Hardware Organization and Design
Linear Prediction Coding of Speech Signal Jun-Won Suh.
Speech Coding Using LPC. What is Speech Coding  Speech coding is the procedure of transforming speech signal into more compact form for Transmission.
Week 2ELE Adaptive Signal Processing 1 STOCHASTIC PROCESSES AND MODELS.
T – Biomedical Signal Processing Chapters
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
1 PATTERN COMPARISON TECHNIQUES Test Pattern:Reference Pattern:
1 Linear Prediction. Outline Windowing LPC Introduction to Vocoders Excitation modeling  Pitch Detection.
Basics of Neural Networks Neural Network Topologies.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Definitions Random Signal Analysis (Review) Discrete Random Signals Random.
Speech Signal Representations I Seminar Speech Recognition 2002 F.R. Verhage.
Adv DSP Spring-2015 Lecture#9 Optimum Filters (Ch:7) Wiener Filters.
Linear Predictive Analysis 主講人:虞台文. Contents Introduction Basic Principles of Linear Predictive Analysis The Autocorrelation Method The Covariance Method.
ECE 5525 Osama Saraireh Fall 2005 Dr. Veton Kepuska
EE513 Audio Signals and Systems
ITU-T G.729 EE8873 Rungsun Munkong March 22, 2004.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Derivation Computational Simplifications Stability Lattice Structures.
0 - 1 © 2007 Texas Instruments Inc, Content developed in partnership with Tel-Aviv University From MATLAB ® and Simulink ® to Real Time with TI DSPs Spectrum.
A Comparison Of Speech Coding With Linear Predictive Coding (LPC) And Code-Excited Linear Predictor Coding (CELP) By: Kendall Khodra Instructor: Dr. Kepuska.
EEL 6586: AUTOMATIC SPEECH PROCESSING Speech Features Lecture Mark D. Skowronski Computational Neuro-Engineering Lab University of Florida February 27,
More On Linear Predictive Analysis
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Normal Equations The Orthogonality Principle Solution of the Normal Equations.
Autoregressive (AR) Spectral Estimation
Chapter 20 Speech Encoding by Parameters 20.1 Linear Predictive Coding (LPC) 20.2 Linear Predictive Vocoder 20.3 Code Excited Linear Prediction (CELP)
Lecture 12: Parametric Signal Modeling XILIANG LUO 2014/11 1.
Review and Summary Box-Jenkins models Stationary Time series AR(p), MA(q), ARMA(p,q)
By Sarita Jondhale 1 Signal preprocessor: “conditions” the speech signal s(n) to new form which is more suitable for the analysis Postprocessor: operate.
EEL 6586: AUTOMATIC SPEECH PROCESSING Speech Features Lecture Mark D. Skowronski Computational Neuro-Engineering Lab University of Florida February 20,
Linear Prediction.
Geology 6600/7600 Signal Analysis 26 Oct 2015 © A.R. Lowry 2015 Last time: Wiener Filtering Digital Wiener Filtering seeks to design a filter h for a linear.
Adv DSP Spring-2015 Lecture#11 Spectrum Estimation Parametric Methods.
PATTERN COMPARISON TECHNIQUES
Figure 11.1 Linear system model for a signal s[n].
Vocoders.
Linear Prediction Simple first- and second-order systems
Linear Prediction.
1 Vocoders. 2 The Channel Vocoder (analyzer) : The channel vocoder employs a bank of bandpass filters,  Each having a bandwidth between 100 HZ and 300.
Modern Spectral Estimation
Microcomputer Systems 2
Linear Predictive Coding Methods
Vocoders.
ESTIMATED INVERSE SYSTEM
Digital Systems: Hardware Organization and Design
Digital Systems: Hardware Organization and Design
Linear Prediction.
Speech Processing Final Project
16. Mean Square Estimation
Presentation transcript:

1 Linear Prediction

2 Linear Prediction (Introduction) : The object of linear prediction is to estimate the output sequence from a linear combination of input samples, past output samples or both : The object of linear prediction is to estimate the output sequence from a linear combination of input samples, past output samples or both : The factors a(i) and b(j) are called predictor coefficients. The factors a(i) and b(j) are called predictor coefficients.

3 Linear Prediction (Introduction) : Many systems of interest to us are describable by a linear, constant-coefficient difference equation : Many systems of interest to us are describable by a linear, constant-coefficient difference equation : If Y(z)/X(z)=H(z), where H(z) is a ratio of polynomials N(z)/D(z), then If Y(z)/X(z)=H(z), where H(z) is a ratio of polynomials N(z)/D(z), then Thus the predicator coefficient given us immediate access to the poles and zeros of H(z). Thus the predicator coefficient given us immediate access to the poles and zeros of H(z).

4 Linear Prediction (Types of System Model) : There are two important variants : There are two important variants : All-pole model (in statistics, autoregressive (AR) model ) : All-pole model (in statistics, autoregressive (AR) model ) : The numerator N(z) is a constant. The numerator N(z) is a constant. All-zero model (in statistics, moving-average (MA) model ) : All-zero model (in statistics, moving-average (MA) model ) : The denominator D(z) is equal to unity. The denominator D(z) is equal to unity. The mixed pole-zero model is called the autoregressive moving-average (ARMA) model. The mixed pole-zero model is called the autoregressive moving-average (ARMA) model.

5 Linear Prediction (Derivation of LP equations) : Given a zero-mean signal y(n), in the AR model : Given a zero-mean signal y(n), in the AR model : The error is : The error is : To derive the predicator we use the orthogonality principle, the principle states that the desired coefficients are those which make the error orthogonal to the samples y(n-1), y(n-2),…, y(n-p). To derive the predicator we use the orthogonality principle, the principle states that the desired coefficients are those which make the error orthogonal to the samples y(n-1), y(n-2),…, y(n-p).

6 Linear Prediction (Derivation of LP equations) : Thus we require that Thus we require that Or, Or, Interchanging the operation of averaging and summing, and representing by summing over n, we have Interchanging the operation of averaging and summing, and representing by summing over n, we have The required predicators are found by solving these equations. The required predicators are found by solving these equations.

7 Linear Prediction (Derivation of LP equations) : The orthogonality principle also states that resulting minimum error is given by The orthogonality principle also states that resulting minimum error is given by Or, Or, We can minimize the error over all time : We can minimize the error over all time : where where

8 Linear Prediction (Applications) : Autocorrelation matching : Autocorrelation matching : We have a signal y(n) with known autocorrelation. We model this with the AR system shown below : We have a signal y(n) with known autocorrelation. We model this with the AR system shown below : σ 1-A(z)

9 Linear Prediction (Order of Linear Prediction) : The choice of predictor order depends on the analysis bandwidth. The rule of thumb is : The choice of predictor order depends on the analysis bandwidth. The rule of thumb is : For a normal vocal tract, there is an average of about one formant per kilohertz of BW. For a normal vocal tract, there is an average of about one formant per kilohertz of BW. One formant require two complex conjugate poles. One formant require two complex conjugate poles. Hence for every formant we require two predicator coefficients, or two coefficients per kilohertz of bandwidth. Hence for every formant we require two predicator coefficients, or two coefficients per kilohertz of bandwidth.

10 Linear Prediction (AR Modeling of Speech Signal) : True Model: True Model: DT Impulse generator G(z) Glottal Filter Uncorrelated Noise generator H(z) Vocal tract Filter R(z) LP Filter Voiced Unvoiced Pitch Gain V U U(n) Voiced Volume velocity s(n) Speech Signal

11 Linear Prediction (AR Modeling of Speech Signal) : Using LP analysis : Using LP analysis : DT Impulse generator White Noise generator All-Pole Filter (AR) Voiced Unvoiced Pitch Gain estimate V U H(z) s(n) Speech Signal

LINEAR PREDICTIVE CODING MODEL FOR SREECH RECOGNITION

The LPC Model Convert this to equality by including an excitation term:

LPC Analysis Equations The prediction error: Error transfer function:

LPC Analysis Equations We seek to minimize the mean squared error signal:

16 Terms of short-term covariance: (*) With this notation, we can write (*) as: A set of P equations, P unknowns

LPC Analysis Equations The minimum mean-squared error can be expressed as:

The Autocorrelation Method w(m): a window zero outside 0≤m≤N-1 The mean squared error is: And:

The Autocorrelation Method

The Autocorrelation Method

The Autocorrelation Method

The Autocorrelation Method

The Autocorrelation Method

The Covariance Method

The Covariance Method The resulting covariance matrix is symmetric, but not Toeplitz, and can be solved efficiently by a set of techniques called Cholesky decomposition

Examples of LPC Analysis

Examples of LPC Analysis

LPC Processor for Speech Recognition

LPC Processor for Speech Recognition Preemphasis: typically a first-order FIR, To spectrally flatten the signal Most widely the following filter is used:

LPC Processor for Speech Recognition Frame Blocking:

LPC Processor for Speech Recognition Hamming Window: Hamming Window: Windowing Windowing Autocorrelation analysis Autocorrelation analysis

LPC Processor for Speech Recognition LPC Analysis, to find LPC coefficients, reflection coefficients (PARCOR), the log area ratio coefficients, the cepstral coefficients, … LPC Analysis, to find LPC coefficients, reflection coefficients (PARCOR), the log area ratio coefficients, the cepstral coefficients, … Durbin’s method Durbin’s method

LPC Processor for Speech Recognition

LPC Processor for Speech Recognition LPC parameter conversion to cepstral coefficients LPC parameter conversion to cepstral coefficients

LPC Processor for Speech Recognition Parameter weighting Parameter weighting Low-order cepstral coefficients are sensitive to overall spectral slope Low-order cepstral coefficients are sensitive to overall spectral slope High-order cepstral coefficients are sensitive to noise High-order cepstral coefficients are sensitive to noise The weighting is done to minimize these sensitivities The weighting is done to minimize these sensitivities

LPC Processor for Speech Recognition

LPC Processor for Speech Recognition Temporal cepstral derivative Temporal cepstral derivative

Typical LPC Analysis Parameters N number of samples in the analysis frame M number of samples shift between frames P LPC analysis order Q dimension of LPC derived cepstral vector K number of frames over which cepstral time derivatives are computed

39 333K Q 10108p 100 (10 msec) 80 (10 msec) 100 (15 msec) M 300 (30 msec) 240 (30 msec) 300 (45 msec) N Typical Values of LPC Analysis Parameters for Speech- Recognition System