Download presentation

Presentation is loading. Please wait.

Published byZander Broyhill Modified over 2 years ago

1
Dr. Gari D. Clifford, University Lecturer & Associate Director, Centre for Doctoral Training in Healthcare Innovation, Institute of Biomedical Engineering, University of Oxford Information Driven Healthcare: Data Visualization & Classification Lecture 1: Introduction & preprocessing Centre for Doctoral Training in Healthcare Innovation

2
The course A practical overview of (a subset of) classifiers and visualization tools Data preparation, PCA, K-means Clustering, KNN Statistics, regression, LDA, logistic regression Neural Networks Gaussian Mixture models, EM Support Vector Machines Labs – try to 1) Classify flowers (classic dataset), … then 2) Predict mortality in the ICU! (... & publish if you do well!)

3
Workload Two lectures each morning Five 4-hour labs (each afternoon) Read one article each eve (optional)

4
Assessment /assignments Class interaction Lab diary – write up notes as you perform investigations – submit lab code (m-file) and Word/OO doc answering the questions at 5pm each day … No paper please! Absolutely no homework!... but you can write a paper afterwards if your results are good!

5
Course texts Ian Nabney, Netlab, Algorithms for Pattern Recognition, in their series Advances in Pattern Recognition. Springer (2001) ISBN Christopher M. Bishop, Pattern Recognition and Machine Learning Springer (2006) ISBN us/um/people/cmbishop/PRML/index.htmhttp://research.microsoft.com/en- us/um/people/cmbishop/PRML/index.htm Press, Teukolsky, Vetterling & Flannery, Numerical Recipes in C, the Art of Scientific Computing, 2nd Edition, Cambrige University Press, [Ch. 2.6, 10.5(p ), 11.0(p ), 15.4(p ), 15.5(p ), 15.6&15.7(p )] Online at L. Tarassenko, A Guide to Neural Computation, John Wiley & Sons (February 1998) Ch. 7 (p77-101) Ian Nabney, Netlab2? - when available! Ian Nabney

6
Syllabus – Week 1 Monday Data exploration [GDC] Lecture 1: (9.30:10.30am) Introduction, probabilities, entropy, preprocessing, normalization, segmenting data (PCA, ICA) Lecture 2: (11am-12pm) Feature extraction, visualization, (K-means, SOM, GTM, Neuroscale). Lab 1 (1-5pm) Preprocessing of data & visualization - segmentation (train, test, evaluation), PCA & K-means with 2 classes Reading for tomorrow: Bishop PRML, Ch4.1 p , Ch4.3.2 p , Ch2.3.7 p ,691, Netlab Ch p Tuesday Clinical Statistics & Classifiers [IS] Lecture 3: (9.30:10.30am) Clinical statistics: t-test, X2 test, Wilcoxon rank sum test, Linear regression, bootstrap, jackknife. Lecture 4: (11am-12pm) Clinical classifiers: LDA, KNN, Logistic Regression Lab 2 (1-5pm) – P-values, statistical testing, LDA, KNN and Logistic regression. Reading for tomorrow: Netlab: Ch p , Ch6 p Wednesday Optimization and Neural Networks [GDC] Lecture 5: (9.30:10.30am) ANNs - RBFs and MLPs - choosing an architecture, balancing the data. Lecture 6: (11am-12pm) Training & optimization, N-fold validation. Lab 3 (1-5pm) Training an MLP to classify flower types and then mortality - partitioning and balancing data, Reading for tomorrow: Netlab: Ch p Thursday Probabilistic Methods [DAC] Lecture 7: (9.30:10.30am) GMM, MCMC, Density Estimation Lecture 8: (11am-12pm) EM, Variation Bayes, missing data Lab 4 (1-5pm) GMM and EM Reading for tomorrow: Bishop: Ch7 p (SVM) Friday Support Vector Machines [CO/GDC] Lecture 9: (9.30:10.30am) SVMs and constrained optimization Lecture 10: (11am-12pm) Wrap-up Lab 5 (1-5pm) Use SVM toolbox and vary 2 parameters for regression & classification (1 class death and then alive), then 2 class.

7
Overview of data for lab You will be given two datasets: 1. A simple dataset for learning – Fishers Iris dataset 2. A complex ICU database (if this works – publish!!!) In each lab you will use dataset 1 to understand the problem, then dataset 2 to see how you can apply this to more challenging data

8
So lets start … what are we doing? Trying to learn classes from data so when we see new data, we can make a good guess concerning its class membership (e.g. is this patient part of the set of people likely to die and if so, can we change his/her treatment) How do we do this? Supervised – use labelled data to train an algorithm. Unsupervised – use heuristics or metrics to look for clusters in data (K-means clustering, KNN, SOMs, GMM, …)

9
Data preprocessing/manipulation Filter data to remove outliers (reject obvious large/small values) Zero-mean, unit variance data if parameters are not in same units! Compress data into lower dimensions to reduce workload or to visualize data relationships Rotate data, or expand into higher dimensions to improve the separation between classes.

10
The curse of dimensionality Richard Bellman (1953) coined the term The Curse of Dimensionality (or Hughes effect) Its the problem caused by the exponential increase in volume associated with adding extra dimensions to a (mathematical) space. Bellman gives the following example: Given 100 evenly-spaced sample points suffice to sample a unit interval with no more than 0.01 distance between points; An equivalent sampling of a 10D unit hypercube with a lattice with a spacing of 0.01 between adjacent points would require 1020 sample points: Therefore, at this spatial sampling resolution, the 10-dimensional hypercube is a factor of 1018 larger than the unit interval. muppet.wikia.com

11
So what does that mean for us? Need to think about how much data we have and how many parameters we use. Rule of thumb: need to have at least 10 training samples of each class per input feature dimension (although this depends on separability of data and can be up to 30 for complex problems and as low as 2-5 for simple problems [*]) So for the Iris dataset – we have 4 measured features on 50 examples of each of the three classes … so we have enough! For ICU data we have 1400 patients, 970 survived and 430 died … so taking the minimum of these we could use up to 43 of the 112 features Generally though you need more data … Or you compress the data into a smaller number of dimensions [*] Thomas G. Van Niel, Tim R. McVicarb and Bisun Datt, On the relationship between training sample size and data dimensionality: Monte Carlo analysis of broadband multi-temporal classification, Remote Sensing of Environment, Volume 98, Issue 4, 30 October 2005, Pages doi: /j.rse

12
Principal Component Analysis (PCA) Standard signal/noise separation method Compress data into lower dimensions to reduce workload or to visualize data relationships Rotate data to improve the separation between classes Also known as Karhunen-Loève (KL) transform or the Hotelling transform or Singular Value Decomposition (SVD) – although SVD is actually a mathematical method of PCA

13
Principal Component Analysis (PCA) A form of Blind source Separation – an observation, X, can be broken down into a mixing matrix, A, and a set of basis functions, Z : X=AZ Second order decorrelation = independence Find a set of orthogonal axes in the data (independence metric = variance) Project data onto these axes to decorrelate Independence is forced onto the data through the orthogonality of axes

14
Two dimensional example Where are the principal components? Hint: axes of maximum variation, and orthogonal

15
Two dimensional example Gives best axis to project minimum RMS error Data becomes sphered or whitened / decorrelated

16
Singular Value Decomposition (SVD) Decompose observation X=AZ into…. X=USV T S is a diagonal matrix of singular values with elements arranged in descending order of magnitude (the singular spectrum) The columns of V are the eigenvectors of C=X T X (the orthogonal subspace … dot(v i,v j )=0 ) … they demix or rotate the data U is the matrix of projections of X onto the eigenvectors of C … the source estimates

17
SVD – matrix algebra Decompose observation X=AZ into…. X=USV T

18
Eigenspectrum of decomposition S = singular matrix … zeros except on the leading diagonal S ij (i=j) are the eigenvalues ½ Placed in order of descending magnitude Correspond to the magnitude of projected data along each eigenvector Eigenvectors are the axes of maximal variation in the data Variance = power (analogous to Fourier components in power spectra) [stem(diag(S).^2)] Eigenspectrum= Plot of eigenvalues

19
SVD: Method for PCA See BSS notes and example at end of presentation

20
SVD for noise/signal separation To perform SVD filtering of a signal, use a truncated SVD decomposition (using the first p eigenvectors) Y=US p V T [Reduce the dimensionality of the data by discarding noise projections S noise =0 Then reconstruct the data with just the signal subsapce] Most of the signal is contained in the first few principal components. Discarding these and projecting back into the original observation space effects a noise-filtering or a noise/signal separation

21
e.g. Imagine a spectral decomposition of the matrix: = xx u1u1 u2u2 1 2 v1v1 v2v2

22
SVD – Dimensionality reduction How exactly is dimension reduction performed? A: Set the smallest singular values to zero: = xx

23
SVD – Dimensionality reduction … note approximation sign xx ~

24
SVD - Dimensionality reduction … and resultant matrix is an approximation using only 3 eigenvectors ~

25
Real ECG data example X X p =US p V T S2S2 X p … p=2 n X p … p=4

26
Recap - PCA Second order decorrelation = independence Find a set of orthogonal axes in the data (independence metric = variance) Project data onto these axes to decorrelate Independence is forced onto the data through the orthogonality of axes Conventional noise / signal separation technique Often used as a method of initializing weights for neural network and other learning algorithms (see Wed lectures).

27
Appendix Worked example (see lecture notes) 4_ICASVDnotes_2009.pdf

28
Worked example

29

30

31

Similar presentations

© 2017 SlidePlayer.com Inc.

All rights reserved.

Ads by Google