Principal Components Analysis (PCA). a technique for finding patterns in data of high dimension.

Slides:



Advertisements
Similar presentations
Krishna Rajan Data Dimensionality Reduction: Introduction to Principal Component Analysis Case Study: Multivariate Analysis of Chemistry-Property data.
Advertisements

Noise & Data Reduction. Paired Sample t Test Data Transformation - Overview From Covariance Matrix to PCA and Dimension Reduction Fourier Analysis - Spectrum.
Covariance Matrix Applications
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Tensors and Component Analysis Musawir Ali. Tensor: Generalization of an n-dimensional array Vector: order-1 tensor Matrix: order-2 tensor Order-3 tensor.
Machine Learning Lecture 8 Data Processing and Representation
PCA + SVD.
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
IDEAL 2005, 6-8 July, Brisbane Multiresolution Analysis of Connectivity Atul Sajjanhar, Deakin University, Australia Guojun Lu, Monash University, Australia.
Lecture 7: Principal component analysis (PCA)
Principal Components Analysis Babak Rasolzadeh Tuesday, 5th December 2006.
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
Principal Component Analysis
Principal Component Analysis
Principal Component Analysis
3D Geometry for Computer Graphics
Pattern Recognition Topic 1: Principle Component Analysis Shapiro chap
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 9(b) Principal Components Analysis Martin Russell.
Face Recognition Using Eigenfaces
CSE 300: Software Reliability Engineering Topics covered: Software metrics and software reliability.
Microarray analysis Algorithms in Computational Biology Spring 2006 Written by Itai Sharon.
Ch. 10: Linear Discriminant Analysis (LDA) based on slides from
Principal Component Analysis. Consider a collection of points.
Techniques for studying correlation and covariance structure
Computer Vision Spring ,-685 Instructor: S. Narasimhan WH 5409 T-R 10:30am – 11:50am Lecture #18.
Principal Component Analysis. Philosophy of PCA Introduced by Pearson (1901) and Hotelling (1933) to describe the variation in a set of multivariate data.
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Empirical Modeling Dongsup Kim Department of Biosystems, KAIST Fall, 2004.
Summarized by Soo-Jin Kim
Principle Component Analysis Presented by: Sabbir Ahmed Roll: FH-227.
Dimensionality Reduction: Principal Components Analysis Optional Reading: Smith, A Tutorial on Principal Components Analysis (linked to class webpage)
Chapter 2 Dimensionality Reduction. Linear Methods
Principal Components Analysis BMTRY 726 3/27/14. Uses Goal: Explain the variability of a set of variables using a “small” set of linear combinations of.
Computer Vision Spring ,-685 Instructor: S. Narasimhan Wean 5403 T-R 3:00pm – 4:20pm Lecture #19.
Additive Data Perturbation: data reconstruction attacks.
What is it? Principal Component Analysis (PCA) is a standard tool in multivariate analysis for examining multidimensional data To reveal patterns between.
Principal Component Analysis Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Classification Course web page: vision.cis.udel.edu/~cv May 12, 2003  Lecture 33.
N– variate Gaussian. Some important characteristics: 1)The pdf of n jointly Gaussian R.V.’s is completely described by means, variances and covariances.
SINGULAR VALUE DECOMPOSITION (SVD)
Descriptive Statistics vs. Factor Analysis Descriptive statistics will inform on the prevalence of a phenomenon, among a given population, captured by.
Principal Component Analysis (PCA). Data Reduction summarization of data with many (p) variables by a smaller set of (k) derived (synthetic, composite)
CSSE463: Image Recognition Day 27 This week This week Today: Applications of PCA Today: Applications of PCA Sunday night: project plans and prelim work.
Rotation matrices 1 Constructing rotation matricesEigenvectors and eigenvalues 0 x y.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Principal Components Analysis. Principal Components Analysis (PCA) A multivariate technique with the central aim of reducing the dimensionality of a multivariate.
EIGENSYSTEMS, SVD, PCA Big Data Seminar, Dedi Gadot, December 14 th, 2014.
Principal Component Analysis Zelin Jia Shengbin Lin 10/20/2015.
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
Feature Extraction 主講人:虞台文.
Presented by: Muhammad Wasif Laeeq (BSIT07-1) Muhammad Aatif Aneeq (BSIT07-15) Shah Rukh (BSIT07-22) Mudasir Abbas (BSIT07-34) Ahmad Mushtaq (BSIT07-45)
CSSE463: Image Recognition Day 25 This week This week Today: Applications of PCA Today: Applications of PCA Sunday night: project plans and prelim work.
Principal Components Analysis ( PCA)
Multivariate statistical methods. Multivariate methods multivariate dataset – group of n objects, m variables (as a rule n>m, if possible). confirmation.
Unsupervised Learning II Feature Extraction
Unsupervised Learning II Feature Extraction
CS246 Linear Algebra Review. A Brief Review of Linear Algebra Vector and a list of numbers Addition Scalar multiplication Dot product Dot product as a.
Principal Component Analysis
Principal Component Analysis (PCA)
COMP 1942 PCA TA: Harry Chan COMP1942.
Lecture: Face Recognition and Feature Reduction
Principal Component Analysis (PCA)
Principal Component Analysis
Techniques for studying correlation and covariance structure
Principal Component Analysis
PCA is “an orthogonal linear transformation that transfers the data to a new coordinate system such that the greatest variance by any projection of the.
Descriptive Statistics vs. Factor Analysis
Feature space tansformation methods
Principal Component Analysis
Presentation transcript:

Principal Components Analysis (PCA)

a technique for finding patterns in data of high dimension

Outline: 1.Eigenvectors and eigenvalues 2.PCA: a)Getting the data b)Centering the data c)Obtaining the covariance matrix d)Performing an eigenvalue decomposition of the covariance matrix e)Choosing components and forming a feature vector f)Deriving the new data set

Outline: 1.Eigenvectors and eigenvalues 2.PCA: a)Getting the data b)Centering the data c)Obtaining the covariance matrix d)Performing an eigenvalue decomposition of the covariance matrix e)Choosing components and forming a feature vector f)Deriving the new data set

Eigenvectors & eigenvalues:

* 23 21

*= 23 21

Eigenvectors & eigenvalues: *= *

Eigenvectors & eigenvalues: *= *=

Eigenvectors & eigenvalues: *= *= = 4 * 3 2

Eigenvectors & eigenvalues: *= *= = 4 * 3 2

Eigenvectors & eigenvalues: *= *= = 4 * 3 2 not an integer multiple of the original vector 4 times the original vector

Eigenvectors & eigenvalues: *= *= = 4 * 3 2 not an integer multiple of the original vector 4 times the original vector transformation matrix eigenvector (this vector and all multiples of it) eigenvalue

Eigenvectors & eigenvalues: *= *= = 4 * 3 2 not an integer multiple of the original vector 4 times the original vector 12 2 (3,2) 8 3 transformation matrix eigenvector (this vector and all multiples of it) eigenvalue

Eigenvectors & eigenvalues: *= *= = 4 * 3 2 not an integer multiple of the original vector 4 times the original vector 12 2 (3,2) 8 3 (12,8) transformation matrix eigenvector (this vector and all multiples of it) eigenvalue

Eigenvectors & eigenvalues: *= *= = 4 * 3 2 not an integer multiple of the original vector 4 times the original vector 12 2 (3,2) 8 3 (12,8) transformation matrix eigenvector (this vector and all multiples of it) eigenvalue

Eigenvectors & eigenvalues: *= *= = 4 * 3 2 not an integer multiple of the original vector 4 times the original vector 12 2 (3,2) 8 3 (12,8) transformation matrix eigenvector (this vector and all multiples of it) eigenvalue

Eigenvectors & eigenvalues: *= *= = 4 * 3 2 not an integer multiple of the original vector 4 times the original vector 12 2 (3,2) 8 3 (12,8) transformation matrix eigenvector (this vector and all multiples of it) eigenvalue

Eigenvectors & eigenvalues: *= *= = 4 * 3 2 not an integer multiple of the original vector 4 times the original vector 12 2 (3,2) 8 3 (12,8) transformation matrix eigenvector (this vector and all multiples of it) eigenvalue

Eigenvectors & eigenvalues: *= *= = 4 * 3 2 not an integer multiple of the original vector 4 times the original vector 12 2 (3,2) 8 3 (12,8) transformation matrix eigenvector (this vector and all multiples of it) eigenvalue

Eigenvectors & eigenvalues: *= *= = 4 * 3 2 not an integer multiple of the original vector 4 times the original vector 12 2 (3,2) 8 3 (12,8) transformation matrix eigenvector (this vector and all multiples of it) eigenvalue

Eigenvectors & eigenvalues: *= *= = 4 * 3 2 not an integer multiple of the original vector 4 times the original vector 12 2 (3,2) 8 3 (12,8) transformation matrix eigenvector (this vector and all multiples of it) eigenvalue

Eigenvectors & eigenvalues: *= *= not an integer multiple of the original vector 4 times the original vector 12 2 (3,2) 8 3 (12,8) transformation matrix eigenvector (this vector and all multiples of it) eigenvalue = 4 *

Eigenvectors & eigenvalues: *= *= not an integer multiple of the original vector 4 times the original vector 12 2 (3,2) 8 3 (12,8) transformation matrix eigenvector (this vector and all multiples of it) eigenvalue = 4 *

Eigenvectors & eigenvalues: *= *= = 4 * 6 4 not an integer multiple of the original vector 4 times the original vector 12 2 (3,2) 8 3 (12,8) transformation matrix eigenvector (this vector and all multiples of it) eigenvalue

Eigenvectors & eigenvalues: *= not an integer multiple of the original vector 4 times the original vector 12 2 (3,2) 8 3 (12,8) transformation matrix eigenvector (this vector and all multiples of it) eigenvalue *= = 4 * 6 4

Outline: 1.Eigenvectors and eigenvalues 2.PCA: a)Getting the data b)Centering the data c)Obtaining the covariance matrix d)Performing an eigenvalue decomposition of the covariance matrix e)Choosing components and forming a feature vector f)Deriving the new data set

Outline: 1.Eigenvectors and eigenvalues 2.PCA: a)Getting the data b)Centering the data c)Obtaining the covariance matrix d)Performing an eigenvalue decomposition of the covariance matrix e)Choosing components and forming a feature vector f)Deriving the new data set

PCA - a technique for identifying patterns in data and expressing the data in such a way as to highlight their similarities and differences - once these patterns are found, the data can be compressed (i.e. the number of dimensions can be reduced) without much loss of information

PCA - a technique for identifying patterns in data and expressing the data in such a way as to highlight their similarities and differences - once these patterns are found, the data can be compressed (i.e. the number of dimensions can be reduced) without much loss of information - example: data of 2 dimensions: xy original data:

PCA - a technique for identifying patterns in data and expressing the data in such a way as to highlight their similarities and differences - once these patterns are found, the data can be compressed (i.e. the number of dimensions can be reduced) without much loss of information - example: data of 2 dimensions: xy xy original data:centered data:

PCA - a technique for identifying patterns in data and expressing the data in such a way as to highlight their similarities and differences - once these patterns are found, the data can be compressed (i.e. the number of dimensions can be reduced) without much loss of information - example: data of 2 dimensions: xy xy original data:centered data:

Covariance matrix: xy x y Eigenvalues of the covariance matrix: Eigenvectors of the covariance matrix: * unit eigenvectors xy xy original data:centered data:

Covariance matrix: xy x y Eigenvalues of the covariance matrix: Eigenvectors of the covariance matrix: * unit eigenvectors xy xy original data:centered data:

Compression and reduced dimensionality: - the eigenvector associated with the highest eigenvalue is the principal component of the dataset; it captures the most significant relationship between the data dimensions - ordering eigenvalues from highest to lowest gives the components in order of significance - if we want, we can ignore the components of lesser significance – this results in a loss of information, but if the eigenvalues are small, the loss will not be great - thus, in a dataset with n dimensions/variables, one may obtain the n eigenvectors & eigenvalues and decide to retain p of them → this results in a final dataset with only p dimensions - feature vector – a vector containing only the eigenvectors representing the dimensions we want to keep – in our example data, 2 choices: - the final dataset is obtained by multiplying the transpose of the feature vector (on the left) with the transposed original dataset - this will give us the original data solely in terms of the vectors we chose or

Our example: a)retain both eigenvectors: b)retain only the first eigenvector / principal component: * data t = new data 2x22x10 * data t = new data 1x22x101x10 Original data, rotated so that the eigenvectors are the axes - no loss of information.

Our example: a)retain both eigenvectors: b)retain only the first eigenvector / principal component: * data t = new data 2x22x10 * data t = new data 1x22x101x10 Original data, rotated so that the eigenvectors are the axes - no loss of information. Only 1 dimension left – we threw away the other axis. Thus: we transformed the data so that is expressed in terms of the patterns, where the patterns are the lines that most closely describe the relationships between the data. Now, the values of the data points tell us exactly where (i.e., above/below) the trend lines the data point sits.

- similar to Cholesky decomposition as one can use both to fully decompose the original covariance matrix - in addition, both produce uncorrelated factors Eigenvalue decomposition: Cholesky decomposition: Eigenvalue: decomposition C = E V E -1 Cholesky decomposition: C = Λ Ψ Λ t c 11 c 21 c 22 * = v1 * e 11 e 21 e 11 e 21 c 11 c 21 c 22 = e 11 e 12 e 21 e 22 v 11 v 22 e 11 e 12 e 21 e 22 c 11 c 21 c 22 = l 11 l 21 l 22 va 11 va 22 l 11 l 21 l 22 ch 1 ch 2 var 1 var 2 va 1 va 2 l 11 l 22 l 21 pc 1 pc 2 var 1 var 2 v 11 v 22 e 11 e 22 e 21 e 12

Thank you for your attention.