CHAPTER 19 Correspondence Analysis From: McCune, B. & J. B. Grace. 2002. Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon.

Slides:



Advertisements
Similar presentations
Tables, Figures, and Equations
Advertisements

CHAPTER 27 Mantel Test From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon
Covariance Matrix Applications
Mutidimensional Data Analysis Growth of big databases requires important data processing.  Need for having methods allowing to extract this information.
An Introduction to Multivariate Analysis
Tensors and Component Analysis Musawir Ali. Tensor: Generalization of an n-dimensional array Vector: order-1 tensor Matrix: order-2 tensor Order-3 tensor.
CHAPTER 24 MRPP (Multi-response Permutation Procedures) and Related Techniques From: McCune, B. & J. B. Grace Analysis of Ecological Communities.
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Lecture 7: Principal component analysis (PCA)
1 Multivariate Statistics ESM 206, 5/17/05. 2 WHAT IS MULTIVARIATE STATISTICS? A collection of techniques to help us understand patterns in and make predictions.
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
Principal Component Analysis
Principal Component Analysis
Linear Transformations
Lecture 6 Ordination Ordination contains a number of techniques to classify data according to predefined standards. The simplest ordination technique is.
Some useful linear algebra. Linearly independent vectors span(V): span of vector space V is all linear combinations of vectors v i, i.e.
From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon
From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon
CHAPTER 22 Reliability of Ordination Results From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach,
From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon
Canonical correlations
Chapter 3 Determinants and Matrices
From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Chapter 7 Data Screening From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon
CHAPTER 30 Structural Equation Modeling From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach,
CHAPTER 20 Detrended Correspondence Analysis From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach,
Chapter 6 Distance Measures From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon
CHAPTER 23 Multivariate Experiments From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon.
CHAPTER 18 Weighted Averaging From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon
From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon
Tables, Figures, and Equations
Techniques for studying correlation and covariance structure
Arithmetic Operations on Matrices. 1. Definition of Matrix 2. Column, Row and Square Matrix 3. Addition and Subtraction of Matrices 4. Multiplying Row.
Summarized by Soo-Jin Kim
Principle Component Analysis Presented by: Sabbir Ahmed Roll: FH-227.
Chapter 2 Dimensionality Reduction. Linear Methods
CHAPTER 26 Discriminant Analysis From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon.
Some matrix stuff.
Principal Component Analysis Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon
Multivariate Data Analysis  G. Quinn, M. Burgman & J. Carey 2003.
N– variate Gaussian. Some important characteristics: 1)The pdf of n jointly Gaussian R.V.’s is completely described by means, variances and covariances.
Principal Component Analysis (PCA). Data Reduction summarization of data with many (p) variables by a smaller set of (k) derived (synthetic, composite)
Principal Components Analysis. Principal Components Analysis (PCA) A multivariate technique with the central aim of reducing the dimensionality of a multivariate.
Eigenvalues The eigenvalue problem is to determine the nontrivial solutions of the equation Ax= x where A is an n-by-n matrix, x is a length n column.
Introduction to Linear Algebra Mark Goldman Emily Mackevicius.
EIGENSYSTEMS, SVD, PCA Big Data Seminar, Dedi Gadot, December 14 th, 2014.
Principle Component Analysis and its use in MA clustering Lecture 12.
Principal Component Analysis (PCA)
Lecture 6 Ordination Ordination contains a number of techniques to classify data according to predefined standards. The simplest ordination technique is.
Discriminant Function Analysis Mechanics. Equations To get our results we’ll have to use those same SSCP matrices as we did with Manova.
Tom.h.wilson Department of Geology and Geography West Virginia University Morgantown, WV.
Principal Components Analysis ( PCA)
Designed by Victor Help you improve MATRICES Let Maths take you Further… Know how to write a Matrix, Know what is Order of Matrices,
Reduced echelon form Matrix equations Null space Range Determinant Invertibility Similar matrices Eigenvalues Eigenvectors Diagonabilty Power.
Introduction to Vectors and Matrices
Principal Component Analysis (PCA)
Systems of Linear Equations
Eigen Decomposition Based on the slides by Mani Thomas and book by Gilbert Strang. Modified and extended by Longin Jan Latecki.
Principal Component Analysis
Introduction to Statistical Methods for Measuring “Omics” and Field Data PCA, PcoA, distance measure, AMOVA.
Recitation: SVD and dimensionality reduction
Eigen Decomposition Based on the slides by Mani Thomas and book by Gilbert Strang. Modified and extended by Longin Jan Latecki.
Matrix Algebra and Random Vectors
Principal Component Analysis (PCA)
Feature space tansformation methods
Introduction to Vectors and Matrices
Principal Component Analysis
Presentation transcript:

CHAPTER 19 Correspondence Analysis From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon Tables, Figures, and Equations

Figure A synthetic data set of eleven species with noiseless hump-shaped responses to an environmental gradient. The gradient was sampled at eleven points (sample units), numbered 1-11.

Figure Comparison of PCA and CA of the data set shown in Figure PCA curves the ends of the gradient in, while CA does not. The vectors indicate the correlations of the environmental gradient with the axis scores.

How it works Axes are rotated simultaneously through species space and sample space with the object of maximizing the correspondence between the two spaces. This produces two sets of linear equations, X and Y, where X = A Y and Y = A' X A is the original data matrix with n rows (henceforth sample units) and p columns (henceforth species). X contains the coordinates for the n sample units on k axes (dimensions). Y contains the coordinates for the p species on k axes. Note that both Y and X refer to the same coordinate system.

The goal is to maximize the correspondence, R, defined as: under the constraints that

Major steps  eigenanalysis approach 1. Calculate weighting coefficients based on the reciprocals of the sample unit totals and the species totals. v contains the sample unit weights, w contains the species weights, a i+ is the total for sample unit i, a +j is the total for species j.

The square roots of these weights are placed in the diagonal of the two matrices V ½ and W ½, which are otherwise filled with zeros. Given n sample units and p species: V ½ has dimensions n  n W ½ has dimensions p  p.

2. Weight the data matrix A by V ½ and W ½ : In other words, This is a simultaneous weighting by row and column totals. The resulting matrix B has n rows and p columns.

3. Calculate a cross-products matrix: S = B'B = V ½ AWA'V ½. The dimensions of S are n  n. The term on the right has dimensions: (n  n)(n  p)(p  p)(p  n)(n  n) Note that S, the cross-products matrix, is a variance- covariance matrix as in PCA except that the cross-products are weighted by the reciprocals of the square roots of the sample unit totals and the species totals.

4. Now find eigenvalues as in PCA. Each eigenvalue (latent root) is a lambda ( ) that solves: │ S - I │ = 0 This is the “characteristic equation.” Note that it is the same as that used in PCA, except for the contents of S.

5. Also find the eigenvectors Y (p  k) and X (n  k) for each of k dimensions: [S - I]x = 0 and [ W ½ A'VAW ½ - I]y = 0 using the same set of in both cases. For each axis there is one and there is one vector x. For every there is one vector y.

6. At this point, we have found X and Y for k dimensions such that: X = A Y n  k n  p p  k and Y = A' X p  k p  n n  k where Y contains the species ordination, A is the original data matrix, and X contains the sample ordination.

Each component or axis can be represented as a linear combination of the original variables. Each eigenvector contains the coefficients for the equation for one axis. For eigenvector 1 (the first column of Y): Score1 x i = y 1 a i1 + y 2 a i y p a ip for entity i

The sample unit scores are scaled by multiplying each element of the SU eigenvectors, X, by where a ++ is the grand total of the matrix A. The species scores are scaled by multiplying each element of the SU eigenvectors, Y, by

Major steps  reciprocal averaging approach 1. Arbitrarily assign scores, x, to the n sample units. The scores position the sample units on an ordination axis.

2. Calculate species scores as weighted averages, where a +j is the total for species j:

3. Calculate new site scores by weighted averaging of the species scores, where a i+ is the total for sample unit i:

4. Center and standardize the site scores so that

5. Check for convergence of the solution. If the site scores are closer than a prescribed tolerance to the site scores of the preceding iteration, then stop. Otherwise, return to step 2.

Figure One dimensional CA ordination of the same data set used in the weighted averaging example in the previous chapter (Fig. 18.1). Scores were standardized to unit variance, then multiplied by 100.

Figure Comparison of 2-D CA (RA), nonmetric multidimensional scaling (NMS), and principal components analysis (PCA) of a data set with known underlying structure. The lines connect sample points along the major environmental gradient. The minor secondary gradient is nearly orthogonal to the major gradient. In the perfect ordination, the points would form a horizontally elongate grid. Inset: the ideal result is a regular 3  10 grid. CA

Table Comparison of correspondence analysis CA, nonmetric multidimensional scaling (NMS), and principal components analysis (PCA) of a data set with known underlying structure.