Md. Zia Uddin Bio-Imaging Lab, Department of Biomedical Engineering Kyung Hee University.

Slides:



Advertisements
Similar presentations
FMRI Methods Lecture 10 – Using natural stimuli. Reductionism Reducing complex things into simpler components Explaining the whole as a sum of its parts.
Advertisements

Independent Component Analysis
Noise & Data Reduction. Paired Sample t Test Data Transformation - Overview From Covariance Matrix to PCA and Dimension Reduction Fourier Analysis - Spectrum.
An Introduction to Multivariate Analysis
Face Recognition Ying Wu Electrical and Computer Engineering Northwestern University, Evanston, IL
Machine Learning Lecture 8 Data Processing and Representation
Dimension reduction (1)
Artifact (artefact) reduction in EEG – and a bit of ERP basics CNC, 19 November 2014 Jakob Heinzle Translational Neuromodeling Unit.
Face Recognition and Biometric Systems
Principal Component Analysis
Independent Component Analysis (ICA)
Application of Statistical Techniques to Neural Data Analysis Aniket Kaloti 03/07/2006.
Dimensional reduction, PCA
Digital Image Processing Chapter 5: Image Restoration.
Face Recognition Using Eigenfaces
Independent Component Analysis (ICA) and Factor Analysis (FA)
Atul Singh Junior Undergraduate CSE, IIT Kanpur.  Dimension reduction is a technique which is used to represent a high dimensional data in a more compact.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
Summarized by Soo-Jin Kim
Dimensionality Reduction: Principal Components Analysis Optional Reading: Smith, A Tutorial on Principal Components Analysis (linked to class webpage)
Chapter 2 Dimensionality Reduction. Linear Methods
ERP DATA ACQUISITION & PREPROCESSING EEG Acquisition: 256 scalp sites; vertex recording reference (Geodesic Sensor Net)..01 Hz to 100 Hz analogue filter;
-1- ICA Based Blind Adaptive MAI Suppression in DS-CDMA Systems Malay Gupta and Balu Santhanam SPCOM Laboratory Department of E.C.E. The University of.
Heart Sound Background Noise Removal Haim Appleboim Biomedical Seminar February 2007.
Eigen Decomposition Based on the slides by Mani Thomas Modified and extended by Longin Jan Latecki.
EEG Classification Using Maximum Noise Fractions and spectral classification Steve Grikschart and Hugo Shi EECS 559 Fall 2005.
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
Basics of Neural Networks Neural Network Topologies.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
ISOMAP TRACKING WITH PARTICLE FILTER Presented by Nikhil Rane.
Lecture 3 BME452 Biomedical Signal Processing 2013 (copyright Ali Işın, 2013) 1 BME452 Biomedical Signal Processing Lecture 3  Signal conditioning.
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
1 CMPB 345: IMAGE PROCESSING DISCRETE TRANSFORM 2.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Principal Component Analysis (PCA)
MACHINE LEARNING 7. Dimensionality Reduction. Dimensionality of input Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Principal components analysis (PCA) as a tool for identifying EEG frequency bands: I. Methodological considerations and preliminary findings Jürgen Kayser,
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
Introduction to Independent Component Analysis Math 285 project Fall 2015 Jingmei Lu Xixi Lu 12/10/2015.
Chapter 13 Discrete Image Transforms
HST.582J/6.555J/16.456J Gari D. Clifford Associate Director, Centre for Doctoral Training, IBME, University of Oxford
Principal Components Analysis ( PCA)
Dimension reduction (1) Overview PCA Factor Analysis Projection persuit ICA.
Unsupervised Learning II Feature Extraction
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Machine Learning Supervised Learning Classification and Regression K-Nearest Neighbor Classification Fisher’s Criteria & Linear Discriminant Analysis Perceptron:
Descriptive Statistics The means for all but the C 3 features exhibit a significant difference between both classes. On the other hand, the variances for.
Medical Image Analysis
Principal Component Analysis (PCA)
Data Transformation: Normalization
Directed Component Analysis
Dimensionality Reduction
LECTURE 11: Advanced Discriminant Analysis
Background on Classification
School of Computer Science & Engineering
Automatic Sleep Stage Classification using a Neural Network Algorithm
Fabien LOTTE, Cuntai GUAN Brain-Computer Interfaces laboratory
Principal Component Analysis (PCA)
Application of Independent Component Analysis (ICA) to Beam Diagnosis
Principal Component Analysis
PCA vs ICA vs LDA.
Blind Source Separation: PCA & ICA
AN ANALYSIS OF TWO COMMON REFERENCE POINTS FOR EEGS
EE513 Audio Signals and Systems
A Fast Fixed-Point Algorithm for Independent Component Analysis
Feature space tansformation methods
Machine Learning for Visual Scene Classification with EEG Data
Principal Component Analysis
Feature Selection in BCIs (section 5 and 6 of Review paper)
Marios Mattheakis and Pavlos Protopapas
Presentation transcript:

Md. Zia Uddin Bio-Imaging Lab, Department of Biomedical Engineering Kyung Hee University

 Abstract  Introduction  Spectral Filtering  Special Filtering  ClassificationContents

 This presentation is about signal processing and machine learning techniques and their applications to BCI.  Overview of general signal processing and classification methods as used in single-trial EEG analysis is given.  For further study, original publications are encouraged. Abstract

 Subject wise experiments ◦ Subject to subject result variances for same kind of experiments.  Session wise experiments ◦ Session to session huge result variability for the same person.  Real time experiments ◦ The system needs to identify the subjects mental state from single trial. ◦ Much more complexity arises. Solution  A session and user brain signature adaptable system is necessary to overcome the subject to subject and session to session huge variability. Why ML for BCI

 Relevant information extraction is difficult because of large dimensional data (i.e., Curse of dimensionality).  Dimensionality has to be reduced keeping the discriminative information and eliminating undiscriminative information.  Most of the classification methods calculate covariance matrix of the data for further feature analysis. Huge covariance matrix is required in the case of large dimensional data.  Thus, Prepropcessiong steps regarding dimensionality reduction is required  In some cases ◦ A priory knowledge is used (e.g., spatial Laplace filter at predefined scalp locations) ◦ Automatic methods (e.g., spatial filters determined by common spatial pattern analysis) Why Preprocessing?

 Common approach is to use digital frequency filter  To consider desired frequency range ◦ Two sequences of poles (a) and zeroes (b) with length n a and n b are necessary that can be calculated by Butterworth or elliptic. ◦ The source signal x is filtered to y as a(1)y(t)=b(1)x(t) + b(2)x(t-1) b( n b )x(t- n b -1) – a(2)y(t-1) a( n a )y(t- n a -1)  Where a and n a are constrained to be 1, is called FIR filter (i.e., considering all zeros).  Advantage of FIR ◦ Produce steeper slopes in between pass and stop band.  In most of the BCI applications, band pass filter is required to consider specific frequency range. Spectral Filter: FIR & IIR

 A good alternative than FIR and IIR is to use temporal Fourier-based filtering in BCI. ◦ A signal switches from temporal to the spectral domain.  The filtered signal is obtained by choosing suitable weighting to the relevant frequency components and applying Inverse Fourier Transformation (IFT).  The short time window determines the frequency resolution Spectral Filter: Fourier-Based Filter

 EEG channels are measured as voltage potential relative to a standard reference (referential recording).  Also, it is possible to record all the channels as voltage difference between the electrode pairs.  From referential EEG, bipolar channels can be obtained by subtracting the respective channels FC4-CP4=(FC4-ref) - (CP4-ref)=FC4 ref -CP4 ref  Reduces the effect of local smearing by computing local gradient.  Focuses on the local activity while contributions of more distant sources are attenuated Spatial Filter: Bipolar Filtering

 The mean of all EEG channels are subtracted from each channel to get the common average reference signals.  Reduces the influence of far field sources but may introduce some undesired spatial smearing  Artifacts of one channel may spread to all other channels. Spatial Filter: Common Average Reference

 More localize filter can be obtained through this.  Laplace signals are obtained by subtracting the average of surrounding electrodes from each individual channel. C4 Lap =C4 ref - ¼(C2 ref + C6 ref + FC4 ref + CP4 ref )  The choice of surrounding channels determine the characteristics of the filter.  Usually, small Laplacians are used (as example given above).  Large Laplacians use neighbors at 20% distance as defined in international system. Spatial Filter: Laplace Filtering

 Represent multidimensional data with fewer number of variables retaining main features of the data.  It is inevitable that by reducing dimensionality some features of the data will be lost. It is hoped that these lost features are comparable with the “noise” and they do not tell much about underlying population.  The method PCA tries to project multidimensional data to a lower dimensional space retaining as much as possible variability of the data.  Its simplicity makes it very popular. But care should be taken in applications. First it should be analyzed if this technique can be applied. Spatial Filter: Principle Component Analysis(1)

 Orthogonal directions of greatest variance in data  Projections along PC1 discriminate the data most along anyone axis  First principal component is the direction of greatest variability (covariance) in the data.  Second is the next orthogonal (uncorrelated) direction of greatest variability ◦ So first remove all the variability along the first component, and then find the next direction of greatest variability  And so on … Spatial Filter: Principle Component Analysis(2) Original Variable A Original Variable B PC 1 PC 2

Spatial Filter: Principle Component Analysis(3)  We can ignore the components of lesser significance.  We do lose some information, but if the eigenvalues are small, we don’t lose much  n dimensions in original data  calculate n eigenvectors and eigenvalues  choose only the first p eigenvectors, based on their eigenvalues  final data set has only p dimensions Subtract the mean Calculate the covariance matrix Calculate the eigenvectors and eigenvalues of the covariance matrix Get data Choosing top components and forming a feature vector Basic Steps Eigenplot

 Problem Definition: ◦ Remove the noise to get VEP in the single trial 29 channels EEG data without ensemble averaging  Technique adopted to solve the Problem: ◦ Selection of principal components as basis for the reconstruction of signal  Methodology ◦ Given signal is divided into an ensemble of signals, for each channel ◦ An ensemble average for each channel is obtained as a reference ◦ Apply PCA to find out the orthonormal eigenvectors which are used as basis for signal approximation ◦ Selection of Principal components as basis by looking at the frequency components present in the “prototype signal” i.e. the averaged signal

Original Signal Single epoch after PCA filtering Reconstructed epoch stacks Filtered signal Original signal Template signal

 Basically ICA is applied for Blind Source Separation (BSS)  Assume an observation (signal) is a linear mix of unknown independent source signals  The mixing (not the signals) is stationary  We have as many observations as unknown sources  To find sources in observations  Need to define a suitable measure of independence  … For example - the cocktail party problem (sources are speakers): Find Z  Formal Statement ◦ N independent sources … Z mn ( M xN ) ◦ linear square mixing … A nn ( N xN ) ◦ produces a set of observations … X mn ( M xN )  ….. X T = AZ T Spatial Filter: Independent Component Analysis(1)

Spatial Filter: Independent Component Analysis(2)  ‘demix’ observations … X T ( N xM ) into Y T = WX T Y T ( N xM )  Z T W ( N xN )  A -1  How do we recover the independent sources?  (We are trying to estimate W  A -1 )  …. We require a measure of independence!

Spatial Filter: Independent Component Analysis(3)  The source signals are mixed by random non orthogonal matrix  JADE algorithm was applied to demix the signals  After reordering and scaling, the demixed signals are very similar to sources.  PCA would fail here as the mixed signals are not orthogonal to each other, which is the key assumption of PCA.  Other ICA algorithms  Infomax  FastICA

Spatial Filter: Independent Component Analysis(4) Applying ICA to single-trial EEG epochs  Data Collection  EEG data were recorded from 31 scalp electrodes  29 placed at locations based on a modified International system  one placed below the right eye (VEOG),  one placed at the left outer canthus (HEOG). All  31 channels were referred to the right mastoid and were digitally sampled for analysis at 256 Hz with a to 100-Hz analog bandpass plus a 50-Hz lowpass filter.  Subjects participated in a 2-hour visual spatial selective attention task in which they were instructed to attend to filled circles flashed in random order in five locations.  Component IC1, generated by blinks  IC4 generated by temporal muscle activity.

Spatial Filter: Independent Component Analysis(5) Applying ICA to single-trial EEG epochs (2)  The scalp maps and power spectra of the 31 independent components derived from target response epochs from a 32-year-old autistic subject.  Blink and eye movement artifact components (IC1 and IC9) had a typical strong low frequency peak.  Temporal muscle artifact components (i.e., ICs 14, 22, 27, and 29) had characteristic focal optima at temporal sites and power plateaus at 20 Hz and higher.

 Next class more classification techniques and some practical examples. Conclusion

Thank you