Multidimensional Data Analysis : the Blind Source Separation problem. Outline : Blind Source Separation Linear mixture model Principal Component Analysis.

Slides:



Advertisements
Similar presentations
Independent Component Analysis
Advertisements

Image Registration  Mapping of Evolution. Registration Goals Assume the correspondences are known Find such f() and g() such that the images are best.
Independent Component Analysis: The Fast ICA algorithm
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The Linear Prediction Model The Autocorrelation Method Levinson and Durbin.
Object Specific Compressed Sensing by minimizing a weighted L2-norm A. Mahalanobis.
Dimension reduction (2) Projection pursuit ICA NCA Partial Least Squares Blais. “The role of the environment in synaptic plasticity…..” (1998) Liao et.
Color Imaging Analysis of Spatio-chromatic Decorrelation for Colour Image Reconstruction Mark S. Drew and Steven Bergner
Dimensionality Reduction PCA -- SVD
Dimension reduction (1)
2008 SIAM Conference on Imaging Science July 7, 2008 Jason A. Palmer
Foreground cleaning in CMB experiments Carlo Baccigalupi, SISSA, Trieste.
FastICA as a LOFAR-EoR Foreground Cleaning Technique Filipe Abdalla and Emma Woodfield University College London with Saleem Zaroubi, Vibor Jelic, Panos.
Independent Component Analysis & Blind Source Separation
EE 290A: Generalized Principal Component Analysis Lecture 6: Iterative Methods for Mixture-Model Segmentation Sastry & Yang © Spring, 2011EE 290A, University.
Independent Component Analysis (ICA)
Component Separation of Polarized Data Application to PLANCK Jonathan Aumont J-F. Macías-Pérez, M. Tristram, D. Santos
Dimensional reduction, PCA
Independent Component Analysis & Blind Source Separation Ata Kaban The University of Birmingham.
Subband-based Independent Component Analysis Y. Qi, P.S. Krishnaprasad, and S.A. Shamma ECE Department University of Maryland, College Park.
Independent Component Analysis (ICA) and Factor Analysis (FA)
ICA Alphan Altinok. Outline  PCA  ICA  Foundation  Ambiguities  Algorithms  Examples  Papers.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation.
Survey on ICA Technical Report, Aapo Hyvärinen, 1999.
Image Representation Gaussian pyramids Laplacian Pyramids
Separating Cosmological B-Modes with FastICA Stivoli F. Baccigalupi C. Maino D. Stompor R. Orsay – 15/09/2005.
Spatial Processes and Image Analysis
The Statistical Properties of Large Scale Structure Alexander Szalay Department of Physics and Astronomy The Johns Hopkins University.
Wavelets, ridgelets, curvelets on the sphere and applications Y. Moudden, J.-L. Starck & P. Abrial Service d’Astrophysique CEA Saclay, France.
Independent Components Analysis with the JADE algorithm
INDEPENDENT COMPONENT ANALYSIS OF TEXTURES based on the article R.Manduchi, J. Portilla, ICA of Textures, The Proc. of the 7 th IEEE Int. Conf. On Comp.
Heart Sound Background Noise Removal Haim Appleboim Biomedical Seminar February 2007.
CMB & Foreground Polarisation CMB 2003 Workshop, Minneapolis Carlo Baccigalupi, SISSA/ISAS.
“A fast method for Underdetermined Sparse Component Analysis (SCA) based on Iterative Detection- Estimation (IDE)” Arash Ali-Amini 1 Massoud BABAIE-ZADEH.
Independent Component Analysis Zhen Wei, Li Jin, Yuxue Jin Department of Statistics Stanford University An Introduction.
Linear Regression Andy Jacobson July 2006 Statistical Anecdotes: Do hospitals make you sick? Student’s story Etymology of “regression”
Joint analysis of Archeops and WMAP observations of the CMB G. Patanchon (University of British Columbia) for the Archeops collaboration.
School of Electrical & Computer Engineering Image Denoising Using Steerable Pyramids Alex Cunningham Ben Clarke Dy narath Eang ECE November 2008.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Cosmic Microwave Background Carlo Baccigalupi, SISSA CMB lectures at TRR33, see the complete program at darkuniverse.uni-hd.de/view/Main/WinterSchoolLecture5.
Estimating the Likelihood of Statistical Models of Natural Image Patches Daniel Zoran ICNC – The Hebrew University of Jerusalem Advisor: Yair Weiss CifAR.
A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent (If f(x) is more complex we usually cannot.
An Introduction to Blind Source Separation Kenny Hild Sept. 19, 2001.
1 MaxEnt CNRS, Paris, France, July 8-13, 2006 “A Minimax Entropy Method for Blind Separation of Dependent Components in Astrophysical Images” Cesar.
EBEx foregrounds and band optimization Carlo Baccigalupi, Radek Stompor.
Principal Manifolds and Probabilistic Subspaces for Visual Recognition Baback Moghaddam TPAMI, June John Galeotti Advanced Perception February 12,
Adventures in Parameter Estimation Jason Dick University of California, Davis.
CSC2515: Lecture 7 (post) Independent Components Analysis, and Autoencoders Geoffrey Hinton.
Principal Component Analysis (PCA)
Blind Component Separation for Polarized Obseravations of the CMB Jonathan Aumont, Juan-Francisco Macias-Perez Rencontres de Moriond 2006 La.
Independent Component Analysis Independent Component Analysis.
Introduction to Independent Component Analysis Math 285 project Fall 2015 Jingmei Lu Xixi Lu 12/10/2015.
An Introduction of Independent Component Analysis (ICA) Xiaoling Wang Jan. 28, 2003.
Chapter 13 Discrete Image Transforms
Dimension reduction (1) Overview PCA Factor Analysis Projection persuit ICA.
CSC2535: Computation in Neural Networks Lecture 7: Independent Components Analysis Geoffrey Hinton.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Planck working group 2.1 diffuse component separation review Paris november 2005.
Biointelligence Laboratory, Seoul National University
LECTURE 10: DISCRIMINANT ANALYSIS
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
Orthogonal Subspace Projection - Matched Filter
Latent Variables, Mixture Models and EM
Application of Independent Component Analysis (ICA) to Beam Diagnosis
PCA vs ICA vs LDA.
A Fast Fixed-Point Algorithm for Independent Component Analysis
Generally Discriminant Analysis
LECTURE 09: DISCRIMINANT ANALYSIS
Independent Factor Analysis
Lecture 16. Classification (II): Practical Considerations
Presentation transcript:

Multidimensional Data Analysis : the Blind Source Separation problem. Outline : Blind Source Separation Linear mixture model Principal Component Analysis : Karhunen-Loeve Basis Gaussianity Mainstream Independent Component Analysis Spectral Matching ICA Diversity and separability Multiscale representations and sparsity Applications : MCA, MMCA, ICAMCA, denoising ICA, etc.

Blind Source Separation Examples : The cocktail party problem

Blind Source Separation Examples : The cocktail party problem Foregrounds in Cosmic Microwave Background imaging –Detector noise –Galactic dust –Synchrotron –Free - Free –Point Sources –Thermal SZ –…

Blind Source Separation Examples : The cocktail party problem Foregrounds in Cosmic Microwave Background imaging Fetal heartbeat estimation

Blind Source Separation Examples : The cocktail party problem Foregrounds in Cosmic Microwave Background imaging Fetal heartbeat estimation Hyperspectral imaging for remote sensing

Blind Source Separation Examples : The cocktail party problem Foregrounds in Cosmic Microwave Background imaging Fetal heartbeat estimation Hyperspectral imaging for remote sensing Differentes views on a single scene obtained with different instruments. COHERENT DATA PROCESSING IS REQUIRED

A simple mixture model linear, static and instantaneous mixture map observed in a given frequency band sums the contributions of various components the contributions of a component to two different detectors differ only in intensity. additive noise all maps are at the same resolution

A simple mixture model 23 GHz 33 GHz 41 GHz 94 GHz Free-free CMB Synchrotron

A simple mixture model 23 GHz 33 GHz 41 GHz 94 GHz Free-free CMB Synchrotron

Processing coherent observations. All component maps are valuable ! How much prior knowledge? – Mixing matrix, component and noise spatial covariances known [Tegmark96, Hobson98]. – Subsets of parameters known. –Blind Source Separation assuming independent processes [Baccigalupi00, Maino92, Kuruoglu03, Delabrouille03]. Huge amounts of high dimensional data. Foreground removal as a source separation problem.

Simplifying assumption The noiseless case : X = AS Can A and S be estimated (up to sign, scale and permutation indeterminacy) from an observed sample of X? Solution is not unique. Prior information is needed.

Estimation of A and S assuming uncorrelated sources and orthogonal mixing matrix. Practically : –Compute the covariance matrix of the data X –Perform an SVD : the estimated eigenvectors are the lines of A Adaptive approximation. Dimensionality reduction; Used for compression and noise reduction. Karhunen-Loeve Transform a.k.a. Principal Component Analysis

Application : original image

Application : KL basis estimated from 12 by 12 patches.

Application : keeping 10 percent of basis elementscompression

Application : keeping 50 percent of basis elementscompression

Still in the noiseless case : X = AS Goal is to find a matrix B such that the entries of Y = BX are independent. Existence/Uniqueness : –Most often, such a decomposition does not exist. –Theorem [Darmois, Linnik 1950] : Let S be a random vector with independent entries of which at most when is Gaussian, and C an invertible matrix. If the entries of Y = CS are independent, then C is almost the identity. Independent Component Analysis Goal will be to find a matrix B such that the entries of Y = BX are as independent as possible.

Independence : Independence or De-correlation? De-correlation measures linear independence: E( (x - E(x)) (y - E(y)) ) = 0 Mutual information or an approximation is used to assess statistical independence:

ICA for gaussian processes? Two linearly mixed independent Gaussian processes do not separate uniquely into two independent components. Need diversity from elsewhere. Algorithms based on non-gaussianity i.e. higher order statistics. Most mainstream ICA techniques: fastICA, Jade, Infomax, etc. Techniques based on the diversity (non proportionality) of variance (energy) profiles in a given representation such as in time, space, Fourier, wavelet : joint diagonalization of covariance matrices, SMICA, etc. Different classes of ICA methods

Example : Jade First, whitening : remove linear dependencies

Example : Jade First, whitening : removes linear dependencies, and normalizes the variance of the projected data. Then, minimize a measure of statistical dependence ie by finding a rotation that minimizes fourth order cross-cumulants.

Example : Jade First, whitening : removes linear dependencies, and normalizes the variance of the projected data. Then, minimize a measure of statistical dependence ie by finding a rotation that minimizes fourth order cross-cumulants.

Application to SZ cluster map restoration Input maps

Application to SZ cluster map restoration Output map Simulated map

ICA of a mixture of independent Gaussian, stationary, colored processes (1) Same mixture model in Fourier Space : X(l) = A S(l) Parameters : the n*n mixing matrix A and the n power spectra D Si (l), for l = 1 to T. Whittle approximation to the likelihood :

ICA of a mixture of independent Gaussian, stationary, colored processes (2) Assuming the spectra are constant over symmetric subintervals, maximizing the likelihood is the same as minimizing: D turns out to be the Kulback Leibler divergence between two gaussian distributions. After minimizing with respect to the source spectra, we are left with a joint diagonality criterion to be minimized wrt B = A -1. There are fast algorithms.

A Gaussian stationary ICA approach to the foreground removal problem? CMB is modeled as a Gaussian Stationary process over the sky. Working with covariance matrices achieves massive data reduction, and the spatial power spectra are the main parameters of interest. Easy extension to the case of noisy mixtures. Connection with Maximum-Likelihood guarantees some kind of optimality. Suggests using EM algorithm. [Snoussi2004, Cardoso2002, Delabrouille2003]

Spectral Matching ICA Back to the case of noisy mixtures : Fit the model spectral covariances : to estimated spectral covariances : by minimizing : with respect to :

Algorithm EM can become very slow, especially as noise parameters become small. Speed up convergence with a few BFGS steps. All /part of the parameters can be estimated.

Estimating component maps Wiener filtering in each frequency band : In case of high SNR, use pseudo inverse :

A variation on SMICA using wavelets Motivations : –Some components, and noise are a priori not stationary. –Galactic components appear correlated. –Emission laws may not be constant over the whole sky. –Incomplete sky coverage.

A variation on SMICA using wavelets Motivations : –Some components, and noise are a priori not stationary. –Galactic components are spatially correlated. –Emission laws may not be constant over the whole sky. –Incomplete sky coverage.

A variation on SMICA using wavelets Motivations : –Some components, and noise are a priori not stationary. –Galactic components appear correlated. –Emission laws may not be constant over the whole sky. –Incomplete sky coverage.

A variation on SMICA using wavelets Motivations : –Some components, and noise are a priori not stationary. –Galactic components appear correlated. –Emission laws may not be constant over the whole sky. –Incomplete sky coverage. Use wavelets to preserve space-scale information.

The à trous wavelet transform shift invariant transform coefficient maps are the same size as the original map isotropic small compact supports of wavelet and scaling functions (B3 spline) Fast algorithms

wSMICA Model structure is not changed : Objective function : Same minimization using EM algorithm and BFGS.

Experiments Simulated Planck HFI observations at high galactic latitudes : Component maps Mixing matrix Planck HFI nominal noise levels

Experiments Simulated Planck HFI observations at high galactic latitudes :

Experiments Contribution of each component to each mixture as a function of spatial frequency :

Results JadewSMICAinitial maps

Results Errors on the estimated emission laws :

Results Rejection rates :where

Multiscale transforms and the quest for sparsity