Central limit theorem - go to web applet. Correlation maps vs. regression maps PNA is a time series of fluctuations in 500 mb heights PNA = 0.25 *

Slides:



Advertisements
Similar presentations
Principal Component Analysis (PCA) or Empirical Orthogonal Functions (EOFs) Arnaud Czaja (SPAT Data analysis lecture Nov. 2011)
Advertisements

Noise & Data Reduction. Paired Sample t Test Data Transformation - Overview From Covariance Matrix to PCA and Dimension Reduction Fourier Analysis - Spectrum.
Dimensionality Reduction PCA -- SVD
PCA + SVD.
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Statistical tools in Climatology René Garreaud
Classical inference and design efficiency Zurich SPM Course 2014
Lecture 7: Principal component analysis (PCA)
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
An introduction to Principal Component Analysis (PCA)
Section 4.2 Fitting Curves and Surfaces by Least Squares.
Principal Component Analysis
The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y: Rank (X Y) =min (rank (X), rank (Y)) A = C S.
Principle Component Analysis What is it? Why use it? –Filter on your data –Gain insight on important processes The PCA Machinery –How to do it –Examples.
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 9(b) Principal Components Analysis Martin Russell.
The Simple Regression Model
TFIDF-space  An obvious way to combine TF-IDF: the coordinate of document in axis is given by  General form of consists of three parts: Local weight.
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Lecture 20 Empirical Orthogonal Functions and Factor Analysis.
Duan Wang Center for Polymer Studies, Boston University Advisor: H. Eugene Stanley.
What is EOF analysis? EOF = Empirical Orthogonal Function Method of finding structures (or patterns) that explain maximum variance in (e.g.) 2D (space-time)
Principal Component Analysis Principles and Application.
Principal Component Analysis. Philosophy of PCA Introduced by Pearson (1901) and Hotelling (1933) to describe the variation in a set of multivariate data.
Chapter 2 Dimensionality Reduction. Linear Methods
Principal Components Analysis BMTRY 726 3/27/14. Uses Goal: Explain the variability of a set of variables using a “small” set of linear combinations of.
Next. A Big Thanks Again Prof. Jason Bohland Quantitative Neuroscience Laboratory Boston University.
Extensions of PCA and Related Tools
Some matrix stuff.
Speaker/ Pei-Ning Kirsten Feng Advisor/ Yu-Heng Tseng
1 The Venzke et al. * Optimal Detection Analysis Jeff Knight * Venzke, S., M. R. Allen, R. T. Sutton and D. P. Rowell, The Atmospheric Response over the.
What is it? Principal Component Analysis (PCA) is a standard tool in multivariate analysis for examining multidimensional data To reveal patterns between.
Principal Component Analysis Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Descriptive Statistics vs. Factor Analysis Descriptive statistics will inform on the prevalence of a phenomenon, among a given population, captured by.
CSE554AlignmentSlide 1 CSE 554 Lecture 8: Alignment Fall 2013.
Modern Navigation Thomas Herring MW 11:00-12:30 Room
Central limit theorem revisited
Contrasts & Statistical Inference
Principal Components Analysis. Principal Components Analysis (PCA) A multivariate technique with the central aim of reducing the dimensionality of a multivariate.
MULTIVARIATE REGRESSION Multivariate Regression; Selection Rules LECTURE 6 Supplementary Readings: Wilks, chapters 6; Bevington, P.R., Robinson, D.K.,
Lecture 12 Factor Analysis.
Principles of the Global Positioning System Lecture 12 Prof. Thomas Herring Room ;
Introduction to Linear Algebra Mark Goldman Emily Mackevicius.
Christina Bonfanti University of Miami- RSMAS MPO 524.
Reduces time complexity: Less computation Reduces space complexity: Less parameters Simpler models are more robust on small datasets More interpretable;
2010/ 11/ 16 Speaker/ Pei-Ning Kirsten Feng Advisor/ Yu-Heng Tseng
EIGENSYSTEMS, SVD, PCA Big Data Seminar, Dedi Gadot, December 14 th, 2014.
PATTERN RECOGNITION : PRINCIPAL COMPONENTS ANALYSIS Richard Brereton
Extratropical Sensitivity to Tropical SST Prashant Sardeshmukh, Joe Barsugli, and Sang-Ik Shin Climate Diagnostics Center.
Intro. ANN & Fuzzy Systems Lecture 16. Classification (II): Practical Considerations.
3 “Products” of Principle Component Analysis
Université d’Ottawa / University of Ottawa 2001 Bio 8100s Applied Multivariate Biostatistics L11.1 Lecture 11: Canonical correlation analysis (CANCOR)
Oceanography 569 Oceanographic Data Analysis Laboratory Kathie Kelly Applied Physics Laboratory 515 Ben Hall IR Bldg class web site: faculty.washington.edu/kellyapl/classes/ocean569_.
Central limit theorem revisited Throw a dice twelve times- the distribution of values is not Gaussian Dice Value Number Of Occurrences.
Principal Components Analysis ( PCA)
Unsupervised Learning II Feature Extraction
Unsupervised Learning II Feature Extraction
PRINCIPAL COMPONENT ANALYSIS(PCA) EOFs and Principle Components; Selection Rules LECTURE 8 Supplementary Readings: Wilks, chapters 9.
CSE 554 Lecture 8: Alignment
Group Meeting R Kirsten Feng.
Fundamentals of regression analysis
Principal Component Analysis
Principal Component Analysis
Contrasts & Statistical Inference
Descriptive Statistics vs. Factor Analysis
Contrasts & Statistical Inference
Principal Component Analysis
Korea Ocean Research & Development Institute, Ansan, Republic of Korea
Contrasts & Statistical Inference
Lecture 16. Classification (II): Practical Considerations
Presentation transcript:

Central limit theorem - go to web applet

Correlation maps vs. regression maps PNA is a time series of fluctuations in 500 mb heights PNA = 0.25 * [ Z(20N,160W) - Z(45N,165W) + Z(55N,115W) - Z(30N,85W) ] index vs. time

PNA - Correlation map (r values of each point with index) PNA - Regression map (meters/std deviation of index) Correlation maps vs. regression maps Correlation maps put each point on ‘equal footing’ Regression maps show magnitude of typical variability PNA is a time series of fluctuations in 500 mb heights PNA = 0.25 * [ Z(20N,160W) - Z(45N,165W) + Z(55N,115W) - Z(30N,85W) ]

Empirical Orthogonal Functions (EOFs): An overview Mathematical technique which decomposes your data matrix into spatial structures (EOFs) and associated amplitude time series (PCs) The EOFs and PCs are constructed to efficiently explain the maximum amount of variance in the data set. By construction, the EOFs are orthogonal to each other, as are the PCs. In general, the majority of the variance in a data set can be explained with just a few EOFs. Provide an ‘objective method’ for finding structure in a data set, but interpretation requires physical facts or intuition. What are the dominant patterns of variabililty in time and space?

Some 2-D Data (X) Singular Value Decomposition (SVD) X = U  V T Eigenanalysis XX T = C; CE =  E 1) Eigenvectors 2) Eigenvalues 3) Principle Components 3 “Products” of Principle Component Analysis

Examples for Today Fake and Real Space- Time Data (X) 1) Eigenvectors – Variations explained in space (MAPS) 2) Eigenvalues - % of Variance explained (spectrum) 3) Principle Components Variations explained in the time (TIMESERIES)

Eigenvectors, Eigenvalues, PC’s Eigenvectors explain variance in one dimension; Principle components explain variance in the other dimension. Each eigenvector has a corresponding principle component. The PAIR define a mode that explains variance. Each eigenvector/PC pair has an associated eigenvalue which relates to how much of the total variance is explained by that mode.

EOF’s and PC’s for geophysical data 1 st EOF is the spatial pattern which explains the most variance of the data in space and time. The 1 st principal component is the time series of the fluctuations of that pattern. 2 nd EOF is the spatial pattern that explains the most of the remaining variance. 2 nd P.C. is the associated time series… EOFs are orthogonal to each other (i.e., e 1.e 2 = 0, where e is vector representing the spatial pattern), and P.C.s are orthogonal to each other (i.e., t 1.t 2 = 0, where t is vector of time series) In general, the majority of the variance in a data set can be explained with just a few EOFs. Go to Joe C.’s photo example….

EOF’s and PC’s for geophysical data By construction, the EOFs are orthogonal to each other, as are the PCs. Provide an ‘objective method’ for finding structure in a data set, but interpretation requires physical facts or intuition.

EOFs: An example based on phony data

EOF 1 PC 1 EOF 2 PC 2

EOF % variance expl. PC 1 EOF % variance expl. PC 2

EOFS: What are they mathematically? Say you have a 2-D data matrix X, where the rows are measurements in time, the columns are measurements in space. The EOFs are eigenvectors of the dispersion matrix X*X T Each eigenvector has an associated eigenvalue which relates to how much of the total variance is explained by that EOF. By solving for the eigenvectors, you have diagonalized the dispersion matrix. This is a coordinate transformation, mapping X*X T into a space where variations are uncorrelated with each other. The PCs and EOFs are related directly through the original data set. The PCs may be obtained by projecting the data set onto the EOFs, and vice versa.

EOF % variance expl. PC 1 EOF % variance expl. PC 2 Eigenvalue Spectrum

EOFs: Example 2 based on contrived data “Data”

EOF % variance expl. PC 1 EOF % variance expl. PC 2 “Data”

EOFs of Real Data: Winter SLP anomalies EOF 1: AO/NAM (23% expl).EOF 2: PNA (13% expl.) EOF 3: non-distinct(10% expl.)

EOF 1 (AO/NAM) EOF 2 (PNA) EOF 3 (?) PC1 (AO/NAM) PC2 (PNA) PC3 (?) EOFs of seal level pressure in the northern hemisphere

EOF significance Each EOF / PC pair comes with an associated eigenvalue The normalized eigenvalues (each eigenvalue divided by the sum of all of the eigenvalues) tells you the percent of variance explained by that EOF / PC pair. Eigenvalues need to be well seperated from each other to be considered distinct modes. First 25 Eigenvalues for DJF SLP

EOF significance: The North Test North et al (1982) provide estimate of error in estimating eigenvalues Requires estimating DOF of the data set. If eigenvalues overlap, those EOFs cannot be considered distinct. Any linear combination of overlapping EOFs is an equally viable structure. First 25 Eigenvalues for DJF SLP Example of overlapping eigenvalues These two just barely overlap. Need physical intuition to help judge.

Validity of EOFs: Questions to ask Is the variance explained more than expected for null hypothesis (red noise, white noise, etc.)? Do we have an a priori reason for expecting this structure? Does it fit with a physical theory? Are the EOFs sensitive to choice of spatial domain? Are the EOFs sensitive to choice of sample? If data set is subdivided (in time), do you still get the same EOFs?

EOFs: Practical Considerations EOFs are easy to calculate, difficult to interpret. There are no hard and fast rules, physical intuition is a must. EOFs are created using linear methods, so they only capture linear relationships. Due to the constraint of orthogonality, EOFs tend to create wave-like structures, even in data sets of pure noise. So pretty… so suggestive… so meaningless. Beware of this. By nature, EOFs give are fixed spatial patterns which only vary in strength and in sign. E.g., the ‘positive’ phase of an EOF looks exactly like the negative phase, just with its sign changed. Many phenomena in the climate system don’t exhibit this kind of symmetry, so EOFs can’t resolve them properly.

Global EOFs (illustrates than domain size matters…)

Global EOFs

Arctic Oscillation in different phases – what are the influences on temperature?

PDO I st EOF of Pacific sea surface temperatures Be careful!!!

What oscillation? What decadal? Be careful!!!