Independent Component Analysis Independent Component Analysis.

Slides:



Advertisements
Similar presentations
Independent Components Analysis
Advertisements

Independent Component Analysis
Independent Component Analysis: The Fast ICA algorithm
Component Analysis (Review)
Dimension reduction (2) Projection pursuit ICA NCA Partial Least Squares Blais. “The role of the environment in synaptic plasticity…..” (1998) Liao et.
Color Imaging Analysis of Spatio-chromatic Decorrelation for Colour Image Reconstruction Mark S. Drew and Steven Bergner
An Information-Maximization Approach to Blind Separation and Blind Deconvolution A.J. Bell and T.J. Sejnowski Computational Modeling of Intelligence (Fri)
1 Maarten De Vos SISTA – SCD - BIOMED K.U.Leuven On the combination of ICA and CPA Maarten De Vos Dimitri Nion Sabine Van Huffel Lieven De Lathauwer.
Microphone Array Post-filter based on Spatially- Correlated Noise Measurements for Distant Speech Recognition Kenichi Kumatani, Disney Research, Pittsburgh.
2008 SIAM Conference on Imaging Science July 7, 2008 Jason A. Palmer
Blind Source Separation of Acoustic Signals Based on Multistage Independent Component Analysis Hiroshi SARUWATARI, Tsuyoki NISHIKAWA, and Kiyohiro SHIKANO.
Variance and covariance M contains the mean Sums of squares General additive models.
REAL-TIME INDEPENDENT COMPONENT ANALYSIS IMPLEMENTATION AND APPLICATIONS By MARCOS DE AZAMBUJA TURQUETI FERMILAB May RTC 2010.
Independent Component Analysis (ICA)
Application of Statistical Techniques to Neural Data Analysis Aniket Kaloti 03/07/2006.
3/24/2006Lecture notes for Speech Communications Multi-channel speech enhancement Chunjian Li DICOM, Aalborg University.
Subband-based Independent Component Analysis Y. Qi, P.S. Krishnaprasad, and S.A. Shamma ECE Department University of Maryland, College Park.
ICA-based Blind and Group-Blind Multiuser Detection.
Independent Component Analysis (ICA) and Factor Analysis (FA)
An Introduction to Independent Component Analysis (ICA) 吳育德 陽明大學放射醫學科學研究所 台北榮總整合性腦功能實驗室.
CS Pattern Recognition Review of Prerequisites in Math and Statistics Prepared by Li Yang Based on Appendix chapters of Pattern Recognition, 4.
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
Estimation and the Kalman Filter David Johnson. The Mean of a Discrete Distribution “I have more legs than average”
ICA Alphan Altinok. Outline  PCA  ICA  Foundation  Ambiguities  Algorithms  Examples  Papers.
Analytical Techniques
Adaptive Signal Processing
Multidimensional Data Analysis : the Blind Source Separation problem. Outline : Blind Source Separation Linear mixture model Principal Component Analysis.
Survey on ICA Technical Report, Aapo Hyvärinen, 1999.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Independent Components Analysis with the JADE algorithm
Independent Component Analysis on Images Instructor: Dr. Longin Jan Latecki Presented by: Bo Han.
Heart Sound Background Noise Removal Haim Appleboim Biomedical Seminar February 2007.
Independent Component Analysis Zhen Wei, Li Jin, Yuxue Jin Department of Statistics Stanford University An Introduction.
Blind Source Separation by Independent Components Analysis Professor Dr. Barrie W. Jervis School of Engineering Sheffield Hallam University England
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent (If f(x) is more complex we usually cannot.
Independent Component Analysis Algorithm for Adaptive Noise Cancelling 적응 잡음 제거를 위한 독립 성분 분석 알고리즘 Hyung-Min Park, Sang-Hoon Oh, and Soo-Young Lee Brain.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
An Introduction to Blind Source Separation Kenny Hild Sept. 19, 2001.
Image cryptosystems based on PottsNICA algorithms Meng-Hong Chen Jiann-Ming Wu Department of Applied Mathematics National Donghwa University.
2010/12/11 Frequency Domain Blind Source Separation Based Noise Suppression to Hearing Aids (Part 2) Presenter: Cian-Bei Hong Advisor: Dr. Yeou-Jiunn Chen.
CSC2515: Lecture 7 (post) Independent Components Analysis, and Autoencoders Geoffrey Hinton.
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
Multi-target Detection in Sensor Networks Xiaoling Wang ECE691, Fall 2003.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Principal Component Analysis (PCA)
Baseband Receiver Receiver Design: Demodulation Matched Filter Correlator Receiver Detection Max. Likelihood Detector Probability of Error.
Introduction to Independent Component Analysis Math 285 project Fall 2015 Jingmei Lu Xixi Lu 12/10/2015.
An Introduction of Independent Component Analysis (ICA) Xiaoling Wang Jan. 28, 2003.
Spatial Covariance Models For Under- Determined Reverberant Audio Source Separation N. Duong, E. Vincent and R. Gribonval METISS project team, IRISA/INRIA,
Spatial vs. Blind Approaches for Speaker Separation: Structural Differences and Beyond Julien Bourgeois RIC/AD.
HST.582J/6.555J/16.456J Gari D. Clifford Associate Director, Centre for Doctoral Training, IBME, University of Oxford
Dimension reduction (1) Overview PCA Factor Analysis Projection persuit ICA.
By: Soroosh Mariooryad Advisor: Dr.Sameti 1 BSS & ICA Speech Recognition - Spring 2008.
CSC2535: Computation in Neural Networks Lecture 7: Independent Components Analysis Geoffrey Hinton.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Lectures 15: Principal Component Analysis (PCA) and
Regression Analysis AGEC 784.
Probability Theory and Parameter Estimation I
LECTURE 11: Advanced Discriminant Analysis
Machine Learning Independent Component Analysis Supervised Learning
LECTURE 10: DISCRIMINANT ANALYSIS
Brain Electrophysiological Signal Processing: Preprocessing
Optimum Passive Beamforming in Relation to Active-Passive Data Fusion
Application of Independent Component Analysis (ICA) to Beam Diagnosis
PCA vs ICA vs LDA.
A Tutorial on Bayesian Speech Feature Enhancement
A Fast Fixed-Point Algorithm for Independent Component Analysis
LECTURE 09: DISCRIMINANT ANALYSIS
Independent Factor Analysis
Presentation transcript:

Independent Component Analysis Independent Component Analysis

Background Hands-free speech recognition system Target Speech Microphone Speech recognition system ? Interference Interference is also observed at microphone. Speech recognition performance significantly degrades. Is it fine tomorrow?

 Goal  Microphone array  Receiver which consists of multiple elements  To enhance target speech or reduce interference  Problem of microphone array processing  A priori information is required.  Directions of arrival of the sound sources  Breaks of target speech for filter adaptation Background (Cont’d)  Realization of high quality hands-free speech interface system

ICA Blind Signal Separation (BSS) or Independent Component Analysis (ICA) is the identification & separation of mixtures of sources with little prior information. Applications include: –Audio Processing –Medical data –Finance –Array processing (beamforming) –Coding … and most applications where Factor Analysis and PCA is currently used. While PCA seeks directions that represents data best in a Σ|x 0 - x| 2 sense, ICA seeks such directions that are most independent from each other. We will concentrate on Time Series separation of Multiple Targets

 Approach taken to estimate source signals only from the observed mixed signals.  Any information about source directions and acoustic conditions is not required.  Independent component Analysis (ICA) is mainly used.  Previous works on ICA J. Cardoso, 1989 C. Jutten, 1990 (Higher-order decorrelation) P. Common, 1994 (define the term “ICA”) A. Bell et al., 1995 (infomax) Blind Source Separation (BSS)

Microphone2 Microphone1 MutuallyIndependent Known ICA-Based BSS Speaker2 Speaker1 Good Morning! Hello! Observed signal1 Observed signal2 Source2 Source1 To estimate source signals No a priori information (unsupervised adaptive filtering)

BSS for Instantaneous mixture Linearly Mixing Process Mixing Matrix Source Observed Separation Process Separated Unmixing Matrix Independent? Cost Function Optimize

Mathematical Formulation s(k)= (s 1 (k),…,s n (k)) T : the vector of n-source signals; x(k)= (x 1 (k),…,x m (k)) T : the vector of m-sensor signals; v (k): the vector of sensor noises. A is the mixing matrix. s(k) x(k)

y(k)= (y 1 (k),…,y m (k)) T : the vector of recovered signals W is the demixing matrix. y(k)=W x(k) Demixing Model Problem: to estimate the source signals (or event-related potentials) by using the sensor signals

Definition Kurtosis is more commonly defined as the fourth cumulant divided by the square of the second cumulant, which is equal to the fourth moment around the mean divided by the square of the variance of the probability distribution minus 3,cumulantvariance which is also known as excess kurtosis. The "minus 3" at the end of this formula is often explained as a correction to make the kurtosis of the normal distribution equal to zero. More generally, if X 1,..., X n are independent random variables all having the same variance, then whereas this identity would not hold if the definition did not include the subtraction of 3.

Various Criterion for ICA Decorrelation –To minimize correlation among signals in multiple time durations Nonlinear function 1 –To minimize higher-order correlation Nonlinear function 2 –To assume p.d.f of sources Separated Signal : Sigmoid function

Cost Function for Nonlinear Function 2 Kullback-Leibler (KL) divergence between and 1. Joint Entropy of 2. Sum of marginal entropy of ・ Minimized when are mutually independent

=

Derivation for Nonlinear Function 2 Nonlinear Function 2 ⇒ To be diagonalized where This can be approximated by Sigmoid Function in speech signal. To update along the negative gradient of

Measures of Non-Gaussianity Kurtotis : gauss=0(sensitive to outliers) Entropy : gauss=largest Neg-entropy : gauss = 0 Approximations where v is a standard gaussian random variable and :

Data Centering & Whitening Centering x = x ‘ – E { x ‘} –But this doesn‘t mean that ICA cannt estimate the mean, but it just simplifies the Alg. –IC‘s are also zero mean because of: E{ s } = W E { x } –After ICA, add W. E { x ‘} to zero mean IC‘s Whitening –We transform the x’s linearly so that the x ~ are white. Its done by EVD. x ~ = (ED -1/2 E T )x = ED -1/2 E T Ax = A ~ s where E {xx ~ } = EDE T So we have to Estimate Orthonormal Matrix A ~ –An orthonormal matrix has n(n-1)/2 degrees of freedom. So for large dim A we have to est only half as much parameters. This greatly simplifies ICA. Reducing dim of data (choosing dominant Eig) while doing whitening also help.

Noisy ICA Model x = As + n A... mxn mixing matrix s... n-dimensional vector of IC‘s n... m-dimensional random noise vector Same assumptions as for noise-free model, if we use measures of nongaussianity which are immune to gaussian noise. So gaussian moments are used as contrast functions. i.e. however, in pre-whitening the effect of noise must be taken in to account: x ~ = ( E{xx T } - Σ) -1/2 x x ~ = Bs + n ~.