EEG Classification Using Maximum Noise Fractions and spectral classification Steve Grikschart and Hugo Shi EECS 559 Fall 2005.

Slides:



Advertisements
Similar presentations
Regression analysis Relating two data matrices/tables to each other Purpose: prediction and interpretation Y-data X-data.
Advertisements

Component Analysis (Review)
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Face Recognition Ying Wu Electrical and Computer Engineering Northwestern University, Evanston, IL
AGC DSP AGC DSP Professor A G Constantinides©1 Modern Spectral Estimation Modern Spectral Estimation is based on a priori assumptions on the manner, the.
Dimension reduction (1)
Nonlinear Regression Ecole Nationale Vétérinaire de Toulouse Didier Concordet ECVPT Workshop April 2011 Can be downloaded at
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
An introduction to Principal Component Analysis (PCA)
Linear Methods for Regression Dept. Computer Science & Engineering, Shanghai Jiao Tong University.
Principal Component Analysis
Dimensional reduction, PCA
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Independent Component Analysis (ICA) and Factor Analysis (FA)
Weighted Median Filters for Complex Array Signal Processing Yinbo Li - Gonzalo R. Arce Department of Electrical and Computer Engineering University of.
Principal Component Analysis Principles and Application.
Techniques for studying correlation and covariance structure
General Linear Model & Classical Inference Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London SPM M/EEGCourse London, May.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
1 Linear Methods for Classification Lecture Notes for CMPUT 466/551 Nilanjan Ray.
Classification (Supervised Clustering) Naomi Altman Nov '06.
With many thanks for slides & images to: FIL Methods group, Virginia Flanagin and Klaas Enno Stephan Dr. Frederike Petzschner Translational Neuromodeling.
SPECTRO-TEMPORAL POST-SMOOTHING IN NMF BASED SINGLE-CHANNEL SOURCE SEPARATION Emad M. Grais and Hakan Erdogan Sabanci University, Istanbul, Turkey  Single-channel.
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
Eigenstructure Methods for Noise Covariance Estimation Olawoye Oyeyele AICIP Group Presentation April 29th, 2003.
Basics of Neural Networks Neural Network Topologies.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Modern Navigation Thomas Herring MW 11:00-12:30 Room
ECE 8443 – Pattern Recognition LECTURE 08: DIMENSIONALITY, PRINCIPAL COMPONENTS ANALYSIS Objectives: Data Considerations Computational Complexity Overfitting.
This supervised learning technique uses Bayes’ rule but is different in philosophy from the well known work of Aitken, Taroni, et al. Bayes’ rule: Pr is.
1 Matrix Algebra and Random Vectors Shyh-Kang Jeng Department of Electrical Engineering/ Graduate Institute of Communication/ Graduate Institute of Networking.
Linear Models for Classification
Robert Engle UCSD and NYU and Robert F. Engle, Econometric Services DYNAMIC CONDITIONAL CORRELATIONS.
Z bigniew Leonowicz, Wroclaw University of Technology Z bigniew Leonowicz, Wroclaw University of Technology, Poland XXIX  IC-SPETO.
Over-fitting and Regularization Chapter 4 textbook Lectures 11 and 12 on amlbook.com.
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
Elements of Pattern Recognition CNS/EE Lecture 5 M. Weber P. Perona.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Normal Equations The Orthogonality Principle Solution of the Normal Equations.
Assignments CS fall Assignment 1 due Generate the in silico data set of 2sin(1.5x)+ N (0,1) with 100 random values of x between.
MACHINE LEARNING 7. Dimensionality Reduction. Dimensionality of input Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Signal Processing in Neuroinformatics EEG Signal Processing Yongnan Ji.
Single Correlator Based UWB Receiver Implementation through Channel Shortening Equalizer By Syed Imtiaz Husain and Jinho Choi School of Electrical Engineering.
Blind Inverse Gamma Correction (Hany Farid, IEEE Trans. Signal Processing, vol. 10 no. 10, October 2001) An article review Merav Kass January 2003.
Yi Jiang MS Thesis 1 Yi Jiang Dept. Of Electrical and Computer Engineering University of Florida, Gainesville, FL 32611, USA Array Signal Processing in.
Linear Classifiers Dept. Computer Science & Engineering, Shanghai Jiao Tong University.
Chapter 15: Classification of Time- Embedded EEG Using Short-Time Principal Component Analysis by Nguyen Duc Thang 5/2009.
Environmental Data Analysis with MatLab 2 nd Edition Lecture 22: Linear Approximations and Non Linear Least Squares.
Unsupervised Learning II Feature Extraction
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
Return to Big Picture Main statistical goals of OODA:
Principal Component Analysis
Background on Classification
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
The general linear model and Statistical Parametric Mapping
Orthogonal Subspace Projection - Matched Filter
CH 5: Multivariate Methods
Modern Spectral Estimation
Ying shen Sse, tongji university Sep. 2016
Techniques for studying correlation and covariance structure
Principal Component Analysis
EE513 Audio Signals and Systems
Matrix Algebra and Random Vectors
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
The general linear model and Statistical Parametric Mapping
Feature space tansformation methods
Generally Discriminant Analysis
Multivariate Methods Berlin Chen
Principal Component Analysis
Multivariate Methods Berlin Chen, 2005 References:
Presentation transcript:

EEG Classification Using Maximum Noise Fractions and spectral classification Steve Grikschart and Hugo Shi EECS 559 Fall 2005

Roadmap  Motivations and background  Available DATA  MNF  Noise covariance estimation  Quadratic Discriminant Analysis  Spectral Discriminant Analysis  Results

Motivations and Background  New capabilities for differently abled persons (i.e. ALS)  Psychomouse!  Divide and conquer approach increases capabilities

EEG Data *  7 subjects, 5 trials of 4 tasks on 2 days  Hz, 6 channels  6 electrodes on electrically linked mastoids  Denote data as 6x2500 matrix, X = (x 1 x 2... x 6 ) *Source:

Data Transformation  Seek a data transformation for easier classification  Optimally using all 6 channel's information  Also exploiting time correlation  Dimension reduction not needed

Maximum Noise Transform (MNF)  Assume signal in additive noise model: X = S + N  Seek a linear combination of data, Xα, that maximizes signal to noise ratio  Express as an optimization problem:

MNF (continued)  When signal and noise components are orthogonal, S T N=N T S=0, equivalently we have:  Generalized Eigenvalue Problem

MNF (continued)  Component with maximum SNR given by top eigenvector  Restrict α ' s by enforcing orthogonality of each solution  SNR of component Xα j given by λ j  Requires estimation of noise covariance N T N  Introduce time correlation by augmenting X matrix

Noise Covariance Estimation  Two basic methods: Differencing: Data – Time-shifted Data Differencing: Data – Time-shifted Data AR fitting: Fit AR to each channel, take residuals AR fitting: Fit AR to each channel, take residuals

Estimation by Differencing  dX = X - X δ, where X δ is a time-shifted version of X  R N = dX T dX = (S+N-S δ -N δ ) T (S+N-S δ -N δ )  Assuming S T N = 0, E[NN δ T ] = 0, S-S δ ≈ 0 then R N = (N-N δ ) T (N-N δ ) ≈ 2N T N = 2Σ N

Estimation by AR fitting  Scalar series vs. vector series  X i (t) = φ 1 X i (t-1) φ q X i (t-q) + ε i (t)  Noise covariance estimated using residuals  Non-linear least squares fit by Gauss- Newton algorithm  Order estimated by AIC (Typical order around 6 * ) (Typical order around 6 * )

QDA But the condition number of the covariance matrix is… e+19

Frequency Domain Classification  Mean signal estimated by averaging across all training data.  Spectral Analysis performed for all training data using Parzen windows, then averaged across all training samples.

Mean estimation

Same day results Misclassifications Correct Classifications 2 task classification 19 4 task classification910

Next day results Misclassifications Correct Classifications 2 task classification task classification3113

Cross person results Misclassifications Correct Classifications 2 task classification task classification3523

Conclusions  This EEG method has promising results but still needs work for acceptable performance  Multi-variate analysis may help  Same day results are good, but not as useful for practical applications