Download presentation

Presentation is loading. Please wait.

Published byTheodore Liggett Modified over 2 years ago

1
Krishna Rajan Data Dimensionality Reduction: Introduction to Principal Component Analysis Case Study: Multivariate Analysis of Chemistry-Property data in Molten Salts C. Suh 1, S. Graduciz 2, M. Gaune-Escard 2, K. Rajan 1 Combinatorial Sciences and Materials Informatics Collaboratory 1 Iowa State University 2 CNRS, Marseilles, France

2
Krishna Rajan From a set of N correlated descriptors, we can derive a set of N uncorrelated descriptors (the principal components). Each principal component (PC) is a suitable linear combination of all the original descriptors. PCA reduces the information dimensionality that is often needed from the vast arrays of data in a way so that there is minimal loss of information ( from Nature Reviews Drug Discovery 1, (2002) : INTEGRATION OF VIRTUAL AND HIGH THROUGHPUT SCREENING Jürgen Bajorath ; and Materials Today; MATERIALS INFORMATICS, K. Rajan, October PRINCIPAL COMPONENT ANALYSIS: PCA

3
Krishna Rajan Functionality 1 = F ( x 1, x 2, x 3, x 4, x 5, x 6, x 7, x 8 ……) Functionality 2 = F ( x 1, x 2, x 3, x 4, x 5, x 6, x 7, x 8 ……) PC 1= A 1 X 1 + A 2 X 2 + A 3 X 3 + A 4 X 4 ……. PC 2 = B 1 X 1 + B 2 X 2 + B 3 X 3 +B 4 X 4 ……. PC 3 = C 1 X 1 + C 2 X 2 + C 3 X 3 + C 4 X 4 ……. X1 = f ( x2) X2 = g( x3) X3= h(x4) ……. I II III …….

4
Krishna Rajan Database of molten salts properties tabulates numerous properties for each chemistry : What can we learn beyond a “search and retrieve” function? Can we find a multivariate correlation (s) among all chemistries and properties? Challenge of reducing the dimensionality of the data set DIMENSIONALITY REDUCTION: Case study

5
Krishna Rajan Principal component analysis (PCA) involves a mathematical procedure that transforms a number of (possibly) correlated variables into a (smaller) number of uncorrelated variables called principal components. The first principal component accounts for as much of the variability in the data as possible, and each succeeding component accounts for as much of the remaining variability as possible.

6
Krishna Rajan Dimensionality Reduction of Molten Salts Data (Janz’s Molten Salts Database:1700 chemistries with 7 variables.) X 1 = f (x 2 ) X 2 = g (x 3 ) X 3 = h (x 4 ) Melting point = F ( x 1, x 2, x 3, x 4, x 5, x 6, x 7, x 8 ……) Density = F ( x 1, x 2, x 3, x 4, x 5, x 6, x 7, x 8 ……) Where x i = molten salt compound chemistries …… …….

7
Krishna Rajan Mathematically, PCA relies on the fact that most of the descriptors are interrelated and these correlations in some instances are high. It results in a rotation of the coordinate system in such a way that the axes show a maximum of variation (covariance) along their directions. This description can be mathematically condensed to a so-called eigenvalue problem. The data manipulation involves decomposition of the data matrix X into two matrices T and P. The two matrices P and T are orthogonal. The matrix P is usually called the loadings matrix, and the matrix T is called the scores matrix. The eigenvectors of the covariance matrix constitute the principal components. The corresponding eigenvalues give a hint to how much "information" is contained in the individual components.

8
Krishna Rajan The loadings can be understood as the weights for each original variable when calculating the principal component. The matrix T contains the original data in a rotated coordinate system. The mathematical analysis involves finding these new “data” matrices T and P. The dimensions of T( ie its rank) that captures all the information of the entire data set of A ( ie # of variables) is far less than that of X ( ideally 2 or 3). One now compresses the N dimensional plot of the data matrix X into 2 or 3 dimensional plot of T and P.

9
Krishna Rajan The first principal component accounts for the maximum variance (eigenvalue) in the original dataset. The second, third ( and higher order) principal components are orthogonal (uncorrelated) to the first and accounts for most of the remaining variance. A new row space is constructed in which to plot the data, where the axes represent the weighted linear combinations of the variables affecting the data. Each of these linear combinations are independent of each other and hence orthogonal. The data when plotted in this new space is essentially a correlation plot, where the position of each data point not only captures all the influences of the variables on that data but also its relative influence compared to the other data. PC 1= A 1 X 1 + A 2 X 2 + A 3 X 3 + A 4 X 4 ……. PC 2 = B 1 X 1 + B 2 X 2 + B 3 X 3 +B 4 X 4 ……. PC 3 = C 1 X 1 + C 2 X 2 + C 3 X 3 + C 4 X 4 …….

10
Krishna Rajan PC1 PC2 PC3 PC4 PC5 …………… Minimal contribution to additional information content beyond higher order principal components.. “Scree” plot helps to identify the # of PCs needed to capture reduced dimensionality NB…depending upon nature of data set, this can be within 2, 3 or higher principal components but still less than the # of variables in original data set Eigenvalue

11
Krishna Rajan Thus the mth PC is orthogonal to all others and has the mth largest variance in the set of PCs. Once the N PCs have been calculated using eigenvalue/ eigenvector matrix operations, only PCs with variances above a critical level are retained (scree test). The M-dimensional principal component space has retained most of the information from the initial N-dimensional descriptor space, by projecting it into orthogonal axes of high variance. The complex tasks of prediction or classification are made easier in this compressed, reduced dimensional space.

12
Krishna Rajan PCA: algorithmic summary

13
Krishna Rajan Dimensionality Reduction of Molten Salts Data (Janz’s Molten Salts Database:1700 instances with 7 variables.) Bivariate representation of the data sets Multivariate (PCA) representation of the data sets

14
Krishna Rajan INTERPRETATIONS OF PRINCIPAL COMPONENT PROJECTIONS Trends in bonding captured along the PC1 axis of scoring plot Correlations between variables captured in loading plot

15
Krishna Rajan To summarize, when we start with a multivariate data matrix PCA analysis permits us to reduce the dimensionality of that data set. This reduction in dimensionality now offers us better opportunities to: Identify the strongest patterns in the data Capture most of the variability of the data by a small fraction of the total set of dimensions Eliminate much of the noise in the data making it beneficial for both data mining and other data analysis algorithms PCA : summary

Similar presentations

© 2017 SlidePlayer.com Inc.

All rights reserved.

Ads by Google