Download presentation
Presentation is loading. Please wait.
1
Independent Component Analysis & Blind Source Separation Ata Kaban The University of Birmingham
2
Overview Today we learn about –The cocktail party problem -- called also ‘blind source separation’ (BSS) –Independent Component Analysis (ICA) for solving BSS –Other applications of ICA / BSS At an intuitive & introductory & practical level
3
A bit like… in the sense of having to find quantities that are not observable directly
4
Signals, joint density time Amplitude S 1 (t) Amplitude S 2 (t) Signals Joint density marginal densities
5
Original signals (hidden sources) s 1 (t), s 2 (t), s 3 (t), s 4 (t), t=1:T
6
The ICA model s1s1 s2s2 s3s3 s4s4 x1x1 x2x2 x3x3 x4x4 a 11 a 12 a 13 a 14 x i (t) = a i1 *s 1 (t) + a i2 *s 2 (t) + a i3 *s 3 (t) + a i4 *s 4 (t) Here, i=1:4. In vector-matrix notation, and dropping index t, this is x = A * s
7
This is recorded by the microphones: a linear mixture of the sources x i (t) = a i1 *s 1 (t) + a i2 *s 2 (t) + a i3 *s 3 (t) + a i4 *s 4 (t)
8
The coctail party problem Called also Blind Source Separation (BSS) problem Ill posed problem, unless assumptions are made! The most common assumption is that source signals are statistically independent. This means that knowing the value of one of them does not give any information about the other. The methods based on this assumption are called Independent Component Analysis methods. These are statistical techniques of decomposing a complex data set into independent parts. It can be shown that under some reasonable conditions, if the ICA assumption holds, then the source signals can be recovered up to permutation and scaling. Determine the source signals, given only the mixtures
9
Recovered signals
10
Some further considerations If we knew the mixing parameters a ij then we would just need to solve a linear system of equations. We know neither a ij nor s i. ICA was initially developed to deal with problems closely related to the coctail party problem Later it became evident that ICA has many other applications too. E.g. from electrical recordings of brain activity from different locations of the scalp (EEG signals) recover underlying components of brain activity
11
Illustration of ICA with 2 signals s1s1 s2s2 x1x1 x2x2 Original sMixed signals a2a2 a1a1 a1a1
12
Illustration of ICA with 2 signals x1x1 x2x2 Step1: Sphering Step2: Rotatation Mixed signals a2a2 a1a1 a1a1
13
Illustration of ICA with 2 signals s1s1 s2s2 x1x1 x2x2 Step1: Sphering Step2: Rotatation Original sMixed signals a2a2 a1a1 a1a1
14
Excluded case There is one case when rotation doesn’t matter. This case cannot be solved by basic ICA. …when both densities are Gaussian Example of non-Gaussian density (-) vs.Gaussian (-.) Seek non-Gaussian sources for two reasons: * identifiability * interestingness: Gaussians are not interesting since the superposition of independent sources tends to be Gaussian
15
Computing the pre-processing steps for ICA 0) Centring = make the signals centred in zero x i x i - E[x i ] for each i 1) Sphering = make the signals uncorrelated. I.e. apply a transform V to x such that Cov(Vx)=I // where Cov(y)=E[yy T ] denotes covariance matrix V=E[xx T ] -1/2 // can be done using ‘sqrtm’ function in MatLab x Vx // for all t (indexes t dropped here) // bold lowercase refers to column vector; bold upper to matrix Scope: to make the remaining computations simpler. It is known that independent variables must be uncorrelated – so this can be fulfilled before proceeding to the full ICA
16
Fixed Point Algorithm Input: X Random init of W Iterate until convergence: Output: W, S where g(.) is derivative of G(.), W is the rotation transform sought Λ is Lagrange multiplier to enforce that W is an orthogonal transform i.e. a rotation Solve by fixed point iterations The effect of Λ is an orthogonal de-correlation Aapo Hyvarinen (97) Computing the rotation step This is based on an the maximisation of an objective function G(.) which contains an approximate non-Gaussianity measure. The overall transform then to take X back to S is (W T V) There are several g(.) options, each will work best in special cases. See FastICA sw / tut for details.
17
Application domains of ICA Blind source separation (Bell&Sejnowski, Te won Lee, Girolami, Hyvarinen, etc.) Image denoising (Hyvarinen) Medical signal processing – fMRI, ECG, EEG (Mackeig) Modelling of the hippocampus and visual cortex (Lorincz, Hyvarinen) Feature extraction, face recognition (Marni Bartlett) Compression, redundancy reduction Watermarking (D Lowe) Clustering (Girolami, Kolenda) Time series analysis (Back, Valpola) Topic extraction (Kolenda, Bingham, Kaban) Scientific Data Mining (Kaban, etc)
18
Image denoising Wiener filtering ICA filtering Noisy image Original image
19
Clustering In multi-variate data search for the direction along of which the projection of the data is maximally non-Gaussian = has the most ‘structure’
20
Blind Separation of Information from Galaxy Spectra
21
Decomposition using Physical Models Decomposition using ICA
22
Summing Up Assumption that the data consists of unknown components –Individual signals in a mix –topics in a text corpus –basis-galaxies Trying to solve the inverse problem: –Observing the superposition only –Recover components –Components often give simpler, clearer view of the data
23
Related resources http://www.cis.hut.fi/projects/ica/cocktail/cocktail_en.cgi Demo and links to further info on ICA. http://www.cis.hut.fi/projects/ica/fastica/code/dlcode.shtml ICA software in MatLab. http://www.cs.helsinki.fi/u/ahyvarin/papers/NN00new.pdf Comprehensive tutorial paper, slightly more technical.
Similar presentations
© 2024 SlidePlayer.com Inc.
All rights reserved.