Heart Sound Background Noise Removal Haim Appleboim Biomedical Seminar February 2007.

Slides:



Advertisements
Similar presentations
Independent Component Analysis
Advertisements

Independent Component Analysis: The Fast ICA algorithm
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The Linear Prediction Model The Autocorrelation Method Levinson and Durbin.
Dimension reduction (2) Projection pursuit ICA NCA Partial Least Squares Blais. “The role of the environment in synaptic plasticity…..” (1998) Liao et.
Lecture 15 Orthogonal Functions Fourier Series. LGA mean daily temperature time series is there a global warming signal?
Experimental Design, Response Surface Analysis, and Optimization
Microphone Array Post-filter based on Spatially- Correlated Noise Measurements for Distant Speech Recognition Kenichi Kumatani, Disney Research, Pittsburgh.
Visual Recognition Tutorial
Independent Component Analysis & Blind Source Separation
REAL-TIME INDEPENDENT COMPONENT ANALYSIS IMPLEMENTATION AND APPLICATIONS By MARCOS DE AZAMBUJA TURQUETI FERMILAB May RTC 2010.
Independent Component Analysis (ICA)
Application of Statistical Techniques to Neural Data Analysis Aniket Kaloti 03/07/2006.
Independent Component Analysis & Blind Source Separation Ata Kaban The University of Birmingham.
3/24/2006Lecture notes for Speech Communications Multi-channel speech enhancement Chunjian Li DICOM, Aalborg University.
ICA-based Blind and Group-Blind Multiuser Detection.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Independent Component Analysis (ICA) and Factor Analysis (FA)
An Introduction to Independent Component Analysis (ICA) 吳育德 陽明大學放射醫學科學研究所 台北榮總整合性腦功能實驗室.
A Quick Practical Guide to PCA and ICA Ted Brookings, UCSB Physics 11/13/06.
Bayesian belief networks 2. PCA and ICA
Some Statistics Stuff (A.K.A. Shamelessly Stolen Stuff)
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
A Unifying Review of Linear Gaussian Models
(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation.
Normalised Least Mean-Square Adaptive Filtering
Survey on ICA Technical Report, Aapo Hyvärinen, 1999.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Independent Components Analysis with the JADE algorithm
-1- ICA Based Blind Adaptive MAI Suppression in DS-CDMA Systems Malay Gupta and Balu Santhanam SPCOM Laboratory Department of E.C.E. The University of.
Independent Component Analysis on Images Instructor: Dr. Longin Jan Latecki Presented by: Bo Han.
Hongyan Li, Huakui Wang, Baojin Xiao College of Information Engineering of Taiyuan University of Technology 8th International Conference on Signal Processing.
Independent Component Analysis Zhen Wei, Li Jin, Yuxue Jin Department of Statistics Stanford University An Introduction.
Blind Separation of Speech Mixtures Vaninirappuputhenpurayil Gopalan REJU School of Electrical and Electronic Engineering Nanyang Technological University.
Blind Source Separation by Independent Components Analysis Professor Dr. Barrie W. Jervis School of Engineering Sheffield Hallam University England
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Basics of Neural Networks Neural Network Topologies.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
A Sparse Non-Parametric Approach for Single Channel Separation of Known Sounds Paris Smaragdis, Madhusudana Shashanka, Bhiksha Raj NIPS 2009.
A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent (If f(x) is more complex we usually cannot.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
An Introduction to Blind Source Separation Kenny Hild Sept. 19, 2001.
2010/12/11 Frequency Domain Blind Source Separation Based Noise Suppression to Hearing Aids (Part 2) Presenter: Cian-Bei Hong Advisor: Dr. Yeou-Jiunn Chen.
CSC2515: Lecture 7 (post) Independent Components Analysis, and Autoencoders Geoffrey Hinton.
3.7 Adaptive filtering Joonas Vanninen Antonio Palomino Alarcos.
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Principal Component Analysis (PCA)
September 28, 2000 Improved Simultaneous Data Reconciliation, Bias Detection and Identification Using Mixed Integer Optimization Methods Presented by:
Independent Component Analysis Independent Component Analysis.
Introduction to Independent Component Analysis Math 285 project Fall 2015 Jingmei Lu Xixi Lu 12/10/2015.
An Introduction of Independent Component Analysis (ICA) Xiaoling Wang Jan. 28, 2003.
Spatial vs. Blind Approaches for Speaker Separation: Structural Differences and Beyond Julien Bourgeois RIC/AD.
HST.582J/6.555J/16.456J Gari D. Clifford Associate Director, Centre for Doctoral Training, IBME, University of Oxford
Dimension reduction (1) Overview PCA Factor Analysis Projection persuit ICA.
CSC2535: Computation in Neural Networks Lecture 7: Independent Components Analysis Geoffrey Hinton.
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
Research Process. Information Theoretic Blind Source Separation with ANFIS and Wavelet Analysis 03 February 2006 서경호.
Blind Extraction of Nonstationary Signal with Four Order Correlation Kurtosis Deconvolution Name: Chong Shan Affiliation: School of Electrical and Information.
LECTURE 11: Advanced Discriminant Analysis
Brain Electrophysiological Signal Processing: Preprocessing
PCA vs ICA vs LDA.
Bayesian belief networks 2. PCA and ICA
Blind Source Separation: PCA & ICA
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Presented by Nagesh Adluru
EE513 Audio Signals and Systems
A Fast Fixed-Point Algorithm for Independent Component Analysis
Unfolding with system identification
Combination of Feature and Channel Compensation (1/2)
Feature Selection in BCIs (section 5 and 6 of Review paper)
Presentation transcript:

Heart Sound Background Noise Removal Haim Appleboim Biomedical Seminar February 2007

Overview Current cardiac monitoring relies on EKG EKG provides full information about the electrical activity but very little information about the mechanical activity Echo provides full info about the mechanical activity Echo can not be used for continuous monitoring Heart Sounds may provide information about mechanical activity

Overview Monitoring heart sounds continuously in non clinical environment is difficult due to background sounds and noise both externally and internally This work introduces a method for removal of internal (and external) background sound such as speech (of the patient)

Overview Heart Sounds are vulnerable to sounds created by the monitored person due to significant overlap in frequency with speech Thus, filters are less effective in removing the noise

Existing solutions Single sensor solutions Multi Sensor solutions Existing Solutions - Time-Varying Wiener filtering - Wavelet - Other filter denoising - Spectral Subtraction Advantages - Easy - Convenient - Good noise removal Disadvantages - Poor noise removal - Inconvenient

A novel approach based on ICA ICA: Independent component analysis is a general method for blind source separation. It has not been used in the context of heart background sound removal We shall demonstrate its superiority over other methods.

ICA (Independent Component Analysis) Sound A Sound B Sound C x = As S1 S m-1 SmSm Each sensor receives a linear mixture of all the signal sources It is required to determine the source signals

ICA definition ICA of a random vector X consists of estimating the following generative model for the data: The independent components are latent variables  They cannot be directly observed The independent components are assumed to have a non-Gaussian distributions The mixing matrix is assumed to be unknown All we observe is the general vector x and we must estimate both A and s using it

Independent Components Computation After estimating the mixing matrix A we can de-mix by computing it’s inverse: Then we obtain the independent components (de-mix) simply by:

Principles of ICA Estimation The fundamental restriction in ICA is that the fundamental components must be non-Gaussian for ICA to be possible The Linear combination of independent variables is more Gaussian than each of the variables To estimate one of the Independent components y let us write: W is One of the rows of A -1 and should be estimated  Linear combination of independent variables is more Gaussian than each of the variables  z T s is more Gaussian than each of the Si  The minimal gaussianity is achieved when it equals to one of the Si We can say that W is a vector which maximizes the nongaussianity of W T X (one of the Independent Components)

FastICA (Hyvärinen, Oja) FastICA learning finds a direction: unit vector w, such that the projection w T x maximizes non-Gaussianity FastICA for one unit computational steps: Choose an initial weight vector If w did not converge go back to 2 FastICA for several units To estimate several independent components the one unit algorithm must run using several units with weight vectors To prevent different vectors converging to the same maxima, after each iteration the outputs must be de-correlated

Spectral ICA: performing ICA on a frequency domain signal (Problem when the Spectrum is complex) Solved using DCT ( Discreet Cosine Transform ) Spectral ICA algorithmic flow Perform DCT on the input data Run FastICA on data after DCT Perform Inverse DCT Spectral ICA

Spectral ICA: Why doe’s it work better? Spectral ICA attempts to separate each frequency component from different sources (ignoring the delays) Since the background sound has overlapping spectrum with the HS but with different magnitudes in each channel, it should be easier to separate and remove

Experimental Setup and Methodology

Quality assessment methods “Human eye” assessment Algorithmic assessment FFT plots Spectrogram plots Diastole Analysis (Noise Removal Quality) Diastole Analysis calculation steps: A = ICA Diastole Peaks average B = NoisySig Diastole Peaks average Noise Removal Quality = B / A Algorithm Quality assessment

Result with Spectral ICA (1) Time Domain HS file name used in this example is ‘GA_Halt_1’, ICA type is ‘Gauss’, Noise file is ‘count_time’

Result with Spectral ICA (2) Time Domain HS file name used in this example is ‘GA_Halt_1’, ICA type is ‘Gauss’, Noise file is ‘television’ Spectral Domain

Result with Spectral ICA (3) Diastole (Quite period) analysis results HS file name used in this example is ‘GA_Halt_1’, Noise file name used is ‘snor_with_pre’ and FastICA type is ‘Gauss’ Wiener Filter FastICA Wiener Filter FastICA Wiener Filter

Cases when ICA doesn't work well Unsuccessful noise removal attempt for a HS with a peak noise caused by microphone movement HS file name used in this example is ‘GA_Halt_1’, FastICA type is ‘Gauss’, Noise file is ‘count_time’

Time vs. Spectral ICA (1) HS file name used in this example is ‘Halt_Supine_1’, Noise file is ‘count_time, FastICA type is ‘Gauss’ Spectral ICA Time ICA Spectral ICA Time ICA

Time vs. Spectral ICA (2) Spectral ICA Time ICA HS file name used in this example is ‘GA_Halt_1’, Noise file name used is ‘snor_with_pre’ and FastICA type is ‘Gauss’ Spectral ICA Time FastICA

Results vs. Nonlinearity ICA nonlinearity = pow3 ICA nonlinearity = tanh ICA nonlinearity = Gauss ICA nonlinearity = skew HS file name used in this example is ‘GA_Normal_1’, Noise file is ‘count_time’

Real Noisy Environment Results HS file name used in this example is ‘HAIM_COUNT_1’, FastICA type is ‘Gauss’

Summery we have introduced a practical method for background noise removal in heart sounds for the purpose of continues heart sounds monitoring The superiority of the proposed method over conventional ones makes it into a practical way of providing high quality heart sound in a noisy environment

Thank You

Principles of ICA Estimation (Measures of nongaussianity) Measures of nongaussianity Kurtosis Negentropy Minimization of mutual information

Kurtosis Kurt (y) = E{y 4 } – 3(E{y 2 }) 2 E{y 2 } = 1  Kurt (y) = E{y 4 } – 3 Y Gaussian  E{y 4 } = 3(E{y 2 }) 2  Kurt (y) = 0 ↓ Therefore kurtosis is 0 for a Gaussian random variable. For most Nongaussian random variables kurtosis ≠ 0 kurtosis can be both positive or negative. Typically nongaussianity is measured by the absolute value of kurtosis.

Negentropy The entropy of a random variable is the degree of information that an observation of the variable gives. Entropy is related to the coding length of the random variable. The more random the variable is, the larger is it’s entropy. For a discrete random variable entropy is defined as: H(Y) = -Σ P(Y=a i ) · log P(Y=a i ) Gaussian variable has the largest entropy among all random variables of equal variance therefore, entropy can be used as a measure of nongaussianity.

Negentropy For a continues random variable differential entropy is defined as: H(y) = -∫f(y)·log(f(y))·dy Negentropy (modified version of differential entropy) is defined as: J(y) = H(y gauss ) – H(y) According to the above definition Negentropy is always non-negative and is zero if and only if y has Gaussian distribution.

Negentropy Practically the estimation of NEGENTROPY is rather difficult. Therefore we usually us some approximations for this purpose

Principles of ICA Estimation (Minimization of mutual information) Definition of mutual information I between m random variables: I(y1,y2, …,ym) = Σ H(yi)-H(y) Mutual Information is a measure of dependence between random variables. It is always non negative and zero if and only if the variables are statistically independent. For an invertible linear transformation y=Wx we can write: I(y 1,y 2,…) = Σ H(y i )-H(x)-log|det(w)|  I(y 1,y 2,…) = Σ C-J(y i )

Principles of ICA Estimation (Minimization of mutual information) Finding the invertible transformation W that minimizes the mutual information is equivalent to fining the direction in which the negentropy J is maximized.