PATTERN RECOGNITION : PRINCIPAL COMPONENTS ANALYSIS Richard Brereton

Slides:



Advertisements
Similar presentations
Krishna Rajan Data Dimensionality Reduction: Introduction to Principal Component Analysis Case Study: Multivariate Analysis of Chemistry-Property data.
Advertisements

3D Geometry for Computer Graphics
PCA for analysis of complex multivariate data. Interpretation of large data tables by PCA In industry, research and finance the amount of data is often.
Regression analysis Relating two data matrices/tables to each other Purpose: prediction and interpretation Y-data X-data.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Pattern Recognition for the Natural Sciences Explorative Data Analysis Principal Component Analysis (PCA) Lutgarde Buydens, IMM, Analytical Chemistry.
Random Sampling and Data Description
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Lecture 7: Principal component analysis (PCA)
1 Multivariate Statistics ESM 206, 5/17/05. 2 WHAT IS MULTIVARIATE STATISTICS? A collection of techniques to help us understand patterns in and make predictions.
Principal Components Analysis Babak Rasolzadeh Tuesday, 5th December 2006.
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
An introduction to Principal Component Analysis (PCA)
Principal Component Analysis
LISA Short Course Series Multivariate Analysis in R Liang (Sally) Shan March 3, 2015 LISA: Multivariate Analysis in RMar. 3, 2015.
Computer Graphics Recitation 5.
Principal component analysis (PCA)
The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y: Rank (X Y) =min (rank (X), rank (Y)) A = C S.
CALIBRATION Prof.Dr.Cevdet Demir
Dimension reduction : PCA and Clustering Christopher Workman Center for Biological Sequence Analysis DTU.
Exploring Microarray data Javier Cabrera. Outline 1.Exploratory Analysis Steps. 2.Microarray Data as Multivariate Data. 3.Dimension Reduction 4.Correlation.
PATTERN RECOGNITION : PRINCIPAL COMPONENTS ANALYSIS Prof.Dr.Cevdet Demir
E.G.M. PetrakisDimensionality Reduction1  Given N vectors in n dims, find the k most important axes to project them  k is user defined (k < n)  Applications:
Principal component analysis (PCA)
Principal component analysis (PCA) Purpose of PCA Covariance and correlation matrices PCA using eigenvalues PCA using singular value decompositions Selection.
Computer Vision Spring ,-685 Instructor: S. Narasimhan WH 5409 T-R 10:30am – 11:50am Lecture #18.
Object Orie’d Data Analysis, Last Time Finished Algebra Review Multivariate Probability Review PCA as an Optimization Problem (Eigen-decomp. gives rotation,
The Tutorial of Principal Component Analysis, Hierarchical Clustering, and Multidimensional Scaling Wenshan Wang.
Dimensionality Reduction: Principal Components Analysis Optional Reading: Smith, A Tutorial on Principal Components Analysis (linked to class webpage)
Chapter 2 Dimensionality Reduction. Linear Methods
Principal Components Analysis BMTRY 726 3/27/14. Uses Goal: Explain the variability of a set of variables using a “small” set of linear combinations of.
CHAPTER 26 Discriminant Analysis From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon.
Computer Vision Spring ,-685 Instructor: S. Narasimhan Wean 5403 T-R 3:00pm – 4:20pm Lecture #19.
Basic concepts in ordination
ArrayCluster: an analytic tool for clustering, data visualization and module finder on gene expression profiles 組員:李祥豪 謝紹陽 江建霖.
1 Dimension Reduction Examples: 1. DNA MICROARRAYS: Khan et al (2001): 4 types of small round blue cell tumors (SRBCT) Neuroblastoma (NB) Rhabdomyosarcoma.
ASCA: analysis of multivariate data from an experimental design, Biosystems Data Analysis group Universiteit van Amsterdam.
Principal Component Analysis Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Chapter 9 Factor Analysis
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Descriptive Statistics vs. Factor Analysis Descriptive statistics will inform on the prevalence of a phenomenon, among a given population, captured by.
A B S T R A C T The study presents the application of selected chemometric techniques to the pollution monitoring dataset, namely, cluster analysis,
Principal Component Analysis (PCA). Data Reduction summarization of data with many (p) variables by a smaller set of (k) derived (synthetic, composite)
PATTERN RECOGNITION : CLUSTERING AND CLASSIFICATION Richard Brereton
Principal Components Analysis. Principal Components Analysis (PCA) A multivariate technique with the central aim of reducing the dimensionality of a multivariate.
EIGENSYSTEMS, SVD, PCA Big Data Seminar, Dedi Gadot, December 14 th, 2014.
Principle Component Analysis and its use in MA clustering Lecture 12.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Principal Component Analysis (PCA)
3D Geometry for Computer Graphics Class 3. 2 Last week - eigendecomposition A We want to learn how the transformation A works:
CSSE463: Image Recognition Day 25 This week This week Today: Applications of PCA Today: Applications of PCA Sunday night: project plans and prelim work.
Principal Components Analysis ( PCA)
Central limit theorem - go to web applet. Correlation maps vs. regression maps PNA is a time series of fluctuations in 500 mb heights PNA = 0.25 *
Multivariate statistical methods. Multivariate methods multivariate dataset – group of n objects, m variables (as a rule n>m, if possible). confirmation.
Unsupervised Learning II Feature Extraction
Unsupervised Learning II Feature Extraction
Self-Modeling Curve Resolution and Constraints Hamid Abdollahi Department of Chemistry, Institute for Advanced Studies in Basic Sciences (IASBS), Zanjan,
Machine Learning Supervised Learning Classification and Regression K-Nearest Neighbor Classification Fisher’s Criteria & Linear Discriminant Analysis Perceptron:
Unsupervised Learning
Principal Component Analysis
LECTURE 11: Advanced Discriminant Analysis
Exploring Microarray data
Principal Component Analysis
Covariance Vs Correlation Matrix
Recitation: SVD and dimensionality reduction
X.1 Principal component analysis
Feature space tansformation methods
Principal Component Analysis
Marios Mattheakis and Pavlos Protopapas
Unsupervised Learning
Presentation transcript:

PATTERN RECOGNITION : PRINCIPAL COMPONENTS ANALYSIS Richard Brereton

NEED FOR PATTERN RECOGNITION Exploratory data analysis e.g. PCA Unsupervised pattern recognition e.g. Cluster analysis Supervised pattern recognition e.g. Classification

Case study Coupled chromatography in HPLC : profile

Time : rows Wavelength : columns MULTIVARIATE DATA

DATA MATRICES The rows do not need to correspond to elution times in chromatography they can be any type of sample Blood sample Wood Chromatograms Samples from a reaction mixture Chromatographic columns

The loadings do not need to correspond to spectral wavelengths they can be any type of sample NMR peak heights Atomic spectroscopy measurements of elements Chromatographic intensities Concentrations of compounds in a mixture Results of chromatographic tests

Return to example of chromatography. Rows : elution times Columns : wavelengths

Chemical factors : X = C.S + E

It would be nice to look at the chemical factors underlying the chromatogram. We can use mathematical methods to do this.

ABSTRACT FACTORS : PRINCIPAL COMPONENTS

X = T. P + E = C. S + E T are called scores: these correspond to elution profile P are called loadings : these correspond to spectra Ideally the “size” of T and P equals the number of compounds in the mixture. This “size” equals the number of principal components, e.g. 1, 2, 3 etc. Each PC has an associated scores vector (column of T), and loadings vector (column of P).

Scores T Data X I J I J A Loadings P A PCA

Hence if the original data matrix is dimensions 30  28 (or I  J) (= 30 elution times and 28 wavelengths - or 30 blood samples and 28 compound concentrations - or 30 chromatographic columns and 28 tests) and if the number of PCs is denoted by A, then the dimensions of T will be 30  A, and the dimensions of P will be A  28.

A major reason for performing PCA is data simplification. Often datasets are very complex, it is possible to make many measurements, but only a few underlying factors. “See the wood from the trees”. Will look at this in more detail later.

SCORES AND LOADINGS HAVE SPECIAL MATHEMATICAL PROPERTIES Scores and loadings are orthogonal. What does this mean? Loadings are normalised. What does this mean?

PCA is an abstract concept. Theory. Non-mathematical Spectrum recorded at different concentrations and several wavelengths; wavelength 6 versus 9 : six spectra.

Each spectrum becomes ONE POINT IN 2 DIMENSIONAL SPACE (2D = 2 wavelengths) Spectra Fall on a straight line which is the FIRST PRINCIPAL COMPONENT The line has a DIRECTION often called the LOADINGS corresponding to the SPECTRAL CHARACTERISTICS Each spectrum has a DISTANCE along the line often called the SCORES corresponding to CONCENTRATION

EXTENSIONS TO THE IDEA 1.Measurement error 2.Several wavelengths 3.Several compounds

Best fit straight line - statistics Two PCs - the second relates to the error around the straight Measurement error

Several wavelengths Now no longer a point in 2 dimensional space. Typical spectrum. Several thousand wavelengths The number of dimensions equals the number of wavelengths. The spectra still fall (roughly) on a straight line. A point in 1000 dimensional space.

Several compounds Two compounds, two wavelengths. A B

RANK AND EIGENVALUE How many PCs describe a dataset? Often unknown How many compounds in a series of mixtures? How many sources of pollution? How many compounds in a reaction mixture? Sometimes just statistical concept. Sometimes mixture of physical and chemical factors, e.g. a reaction mixture : compounds, temperature etc.

EVERY PRINCIPAL COMPONENT HAS A CORRESPONDING EIGENVALUE The eigenvalue equals the sum of squares of the scores vector for each PC. The more important the PC the bigger the eigenvalue. The sum of squares of the eigenvalues of a matrix should never exceed that of the original matrix. The sum of squares of all significant PCs should approximate to that of the original matrix.

RESIDUAL SUM OF SQUARES : decreases as the number of eigenvalues increases. Log eigenvalue versus component number. Cut off?

SEVERAL OTHER APPROACHES FOR THE DETERMINATION OF NUMBER OF EIGENVALUES.

SUMMARY SO FAR PCA Principal components – how many? Scores Loadings Eigenvalues

GRAPHIC DISPLAY OF PCS SCORES PLOT PC2 VERSUS PC1

SCORES AGAINST TIME PC1 AND PC2 VERSUS TIME

LOADINGS PLOT PC2 VERSUS PC1

FOR REFERENCE : pure spectra

LOADINGS AGAINST WAVELENGTH PC1 AND PC2 VERSUS WAVELENGTH

BIPLOTS : SUPERIMPOSING SCORES AND LOADINGS PLOTS

MANY OTHER PLOTS Not only PC2 versus 1, also PC3 versus 1, PC3 versus 2 etc. 3D PC plots, 3 axes, rotation etc. Loadings and scores sometimes presented as bar graphs, not always a sequential meaning. Plots of eigenvalues against component number

DATA SCALING AND PREPROCESSING Influences appearance of plots Column centring – common in traditional statistics Standardisation of columns – subtract mean and divide by standard deviation. If data of different types or absolute scales this is an essential technique Row scaling – to constant total

ANOTHER EXAMPLE Grouping of elements from fundamental properties using PCA.

Step 1 : standardise the data. Why? On different scales.

PERFORM PCA : Choose the first two PCs Scores plot

Loadings plot

SUMMARY Many types of plot from PCA. Interpretation of the plots. Preprocessing important.