Cornea Data Main Point: OODA Beyond FDA Recall Interplay: Object Space  Descriptor Space.

Slides:



Advertisements
Similar presentations
Object Orie’d Data Analysis, Last Time •Clustering –Quantify with Cluster Index –Simple 1-d examples –Local mininizers –Impact of outliers •SigClust –When.
Advertisements

Eigen Decomposition and Singular Value Decomposition
3D Geometry for Computer Graphics
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Object Orie’d Data Analysis, Last Time Finished NCI 60 Data Started detailed look at PCA Reviewed linear algebra Today: More linear algebra Multivariate.
Participant Presentations Please Sign Up: Name (Onyen is fine, or …) Are You ENRolled? Tentative Title (???? Is OK) When: Next Week, Early, Oct.,
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
An introduction to Principal Component Analysis (PCA)
3D Geometry for Computer Graphics
Principal component analysis (PCA)
Dimensional reduction, PCA
The UNIVERSITY of Kansas EECS 800 Research Seminar Mining Biological Data Instructor: Luke Huan Fall, 2006.
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
3D Geometry for Computer Graphics
Exploring Microarray data Javier Cabrera. Outline 1.Exploratory Analysis Steps. 2.Microarray Data as Multivariate Data. 3.Dimension Reduction 4.Correlation.
Principal component analysis (PCA) Purpose of PCA Covariance and correlation matrices PCA using eigenvalues PCA using singular value decompositions Selection.
Techniques for studying correlation and covariance structure
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka Virginia de Sa (UCSD) Cogsci 108F Linear.
Statistics – O. R. 892 Object Oriented Data Analysis J. S. Marron Dept. of Statistics and Operations Research University of North Carolina.
Object Orie’d Data Analysis, Last Time Finished Algebra Review Multivariate Probability Review PCA as an Optimization Problem (Eigen-decomp. gives rotation,
Detailed Look at PCA Three Important (& Interesting) Viewpoints: 1. Mathematics 2. Numerics 3. Statistics 1 st : Review Linear Alg. and Multivar. Prob.
Chapter 2 Dimensionality Reduction. Linear Methods
Probability of Error Feature vectors typically have dimensions greater than 50. Classification accuracy depends upon the dimensionality and the amount.
Principal Components Analysis BMTRY 726 3/27/14. Uses Goal: Explain the variability of a set of variables using a “small” set of linear combinations of.
Linear Algebra Review 1 CS479/679 Pattern Recognition Dr. George Bebis.
Object Orie’d Data Analysis, Last Time Gene Cell Cycle Data Microarrays and HDLSS visualization DWD bias adjustment NCI 60 Data Today: More NCI 60 Data.
Object Orie’d Data Analysis, Last Time
Next. A Big Thanks Again Prof. Jason Bohland Quantitative Neuroscience Laboratory Boston University.
Robust PCA Robust PCA 3: Spherical PCA. Robust PCA.
Object Orie’d Data Analysis, Last Time Primal – Dual PCA vs. SVD (not comparable) Vectors (discrete) vs. Functions (contin ’ s) PCA for shapes – Corpus.
N– variate Gaussian. Some important characteristics: 1)The pdf of n jointly Gaussian R.V.’s is completely described by means, variances and covariances.
Object Orie’d Data Analysis, Last Time Discrimination for manifold data (Sen) –Simple Tangent plane SVM –Iterated TANgent plane SVM –Manifold SVM Interesting.
Object Orie’d Data Analysis, Last Time Cornea Data –Images (on the disk) as data objects –Zernike basis representations Outliers in PCA (have major influence)
Descriptive Statistics vs. Factor Analysis Descriptive statistics will inform on the prevalence of a phenomenon, among a given population, captured by.
Object Orie’d Data Analysis, Last Time Gene Cell Cycle Data Microarrays and HDLSS visualization DWD bias adjustment NCI 60 Data Today: Detailed (math ’
Techniques for studying correlation and covariance structure Principal Components Analysis (PCA) Factor Analysis.
Object Orie’d Data Analysis, Last Time
SWISS Score Nice Graphical Introduction:. SWISS Score Toy Examples (2-d): Which are “More Clustered?”
Discriminant Analysis
Linear Subspace Transforms PCA, Karhunen- Loeve, Hotelling C306, 2000.
Introduction to Linear Algebra Mark Goldman Emily Mackevicius.
Participant Presentations Please Sign Up: Name (Onyen is fine, or …) Are You ENRolled? Tentative Title (???? Is OK) When: Thurs., Early, Oct., Nov.,
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
Instructor: Mircea Nicolescu Lecture 8 CS 485 / 685 Computer Vision.
Object Orie’d Data Analysis, Last Time
Object Orie’d Data Analysis, Last Time PCA Redistribution of Energy - ANOVA PCA Data Representation PCA Simulation Alternate PCA Computation Primal – Dual.
1 UNC, Stat & OR Hailuoto Workshop Object Oriented Data Analysis, I J. S. Marron Dept. of Statistics and Operations Research, University of North Carolina.
Object Orie’d Data Analysis, Last Time PCA Redistribution of Energy - ANOVA PCA Data Representation PCA Simulation Alternate PCA Computation Primal – Dual.
PCA as Optimization (Cont.) Recall Toy Example Empirical (Sample) EigenVectors Theoretical Distribution & Eigenvectors Different!
Feature Extraction 主講人:虞台文.
GWAS Data Analysis. L1 PCA Challenge: L1 Projections Hard to Interpret (i.e. Little Data Insight) Solution: 1)Compute PC Directions Using L1 2)Compute.
Participant Presentations Draft Schedule Now on Course Web Page: When You Present: Please Load Talk on Classroom.
PCA Data Represent ’ n (Cont.). PCA Simulation Idea: given Mean Vector Eigenvectors Eigenvalues Simulate data from Corresponding Normal Distribution.
Multivariate statistical methods. Multivariate methods multivariate dataset – group of n objects, m variables (as a rule n>m, if possible). confirmation.
Object Orie’d Data Analysis, Last Time Finished NCI 60 Data Linear Algebra Review Multivariate Probability Review PCA as an Optimization Problem (Eigen-decomp.
Object Orie’d Data Analysis, Last Time DiProPerm Test –Direction – Projection – Permutation –HDLSS hypothesis testing –NCI 60 Data –Particulate Matter.
Return to Big Picture Main statistical goals of OODA:
Object Orie’d Data Analysis, Last Time
Exploring Microarray data
University of Ioannina
LECTURE 10: DISCRIMINANT ANALYSIS
Functional Data Analysis
Statistics – O. R. 881 Object Oriented Data Analysis
Participant Presentations
Probabilistic Models with Latent Variables
Feature space tansformation methods
LECTURE 09: DISCRIMINANT ANALYSIS
Principal Component Analysis
Participant Presentations
Presentation transcript:

Cornea Data Main Point: OODA Beyond FDA Recall Interplay: Object Space  Descriptor Space

Robust PCA What is multivariate median? There are several! ( “ median ” generalizes in different ways) i.Coordinate-wise median Often worst Not rotation invariant (2-d data uniform on “ L ” ) Can lie on convex hull of data (same example) Thus poor notion of “ center ”

Robust PCA

M-estimate (cont.): “ Slide sphere around until mean (of projected data) is at center ”

Robust PCA Robust PCA 3: Spherical PCA Idea: use “ projection to sphere ” idea from M-estimation In particular project data to centered sphere “ Hot Dog ” of data becomes “ Ice Caps ” Easily found by PCA (on proj ’ d data) Outliers pulled in to reduce influence Radius of sphere unimportant

Robust PCA Spherical PCA for Toy Example: Curve Data With an Outlier First recall Conventional PCA

Robust PCA Spherical PCA for Toy Example: Now do Spherical PCA Better result?

Robust PCA Rescale Coords Unscale Coords Spherical PCA

Robust PCA Elliptical PCA for cornea data: Original PC2, Elliptical PC2

Big Picture View of PCA Alternate Viewpoint: Gaussian Likelihood When data are multivariate Gaussian PCA finds major axes of ellipt ’ al contours of Probability Density Maximum Likelihood Estimate Mistaken idea: PCA only useful for Gaussian data

Big Picture View of PCA Raw Cornea Data: Data – Median (Data – Median) MAD

Big Picture View of PCA Check for Gaussian dist ’ n: Stand ’ zed Parallel Coord. Plot E.g. Cornea data (recall image view of data) Several data points > 20 “ s.d.s ” from the center Distribution clearly not Gaussian Strong kurtosis ( “ heavy tailed ” ) But PCA still gave strong insights

GWAS Data Analysis

Genome Wide Association Study (GWAS) Cystic Fibrosis Study: Wright et al (2011) Interesting Feature: Some Subjects are Close Relatives (e.g. ~half SNPs are same)

GWAS Data Analysis PCA View Clear Ethnic Groups And Several Outliers! Eliminate With Spherical PCA?

GWAS Data Analysis Spherical PCA Looks Same?!? What is going on?

GWAS Data Analysis

L1 PCA Challenge: L1 Projections Hard to Interpret 2-d Toy Example Note Outlier

L1 PCA Conventional L2 PCA Outlier Pulls Off PC1 Direction

L1 PCA Much Better PC1 Direction

L1 PCA Much Better PC1 Direction But Very Strange Projections (i.e. Little Data Insight)

L1 PCA Reason: SVD Rotation Before L1 Computation Note: L1 Methods Not Rotation Invariant

L1 PCA Challenge: L1 Projections Hard to Interpret (i.e. Little Data Insight) Solution: 1)Compute PC Directions Using L1 2)Compute Projections (Scores) Using L2 Called “Visual L1 PCA” (VL1PCA) Recent work with Yihui Zhou

L1 PCA VL1 PCA Excellent PC Directions Interpretable Scores

VL1 PCA 10-d Toy Example: Parabolas From Before With Two Outliers

VL1 PCA 10-d Toy Example: L2PCA Strongly Impacted (In All Components)

VL1 PCA 10-d Toy Example: SCPA & VL1PCA Nicely Robust (Vl1PCA Slightly Better)

VL1 PCA 10-d Toy Example: L1PCA Hard To Interpret

GWAS Data L1 PCA No Longer Feels Outliers But Still Highlights Individuals

GWAS Data VL1 PCA Best Focus On Ethnic Groups

GWAS Data VL1 PCA Best Focus On Ethnic Groups Seems To Find Unlabelled Clusters

Detailed Look at PCA Now Study “Folklore” More Carefully BackGround History Underpinnings (Mathematical & Computational) Good Overall Reference: Jolliffe (2002)

PCA: Rediscovery – Renaming Statistics: Principal Component Analysis (PCA) Social Sciences: Factor Analysis (PCA is a subset) Probability / Electrical Eng: Karhunen – Loeve expansion Applied Mathematics: Proper Orthogonal Decomposition (POD) Geo-Sciences: Empirical Orthogonal Functions (EOF)

An Interesting Historical Note The 1 st (?) application of PCA to Functional Data Analysis: Rao (1958) 1 st Paper with “Curves as Data Objects” viewpoint

Detailed Look at PCA Three Important (& Interesting) Viewpoints: 1. Mathematics 2. Numerics 3. Statistics Goal: Study Interrelationships

Detailed Look at PCA Three Important (& Interesting) Viewpoints: 1. Mathematics 2. Numerics 3. Statistics 1 st : Review Linear Alg. and Multivar. Prob.

Review of Linear Algebra

Review of Linear Algebra (Cont.) SVD Full Representation: = Graphics Display Assumes

Review of Linear Algebra (Cont.) SVD Full Representation: = Full Rank Basis Matrix

Review of Linear Algebra (Cont.) SVD Full Representation: = Full Rank Basis Matrix All 0s in Bottom (and off diagonal)

Review of Linear Algebra (Cont.) SVD Reduced Representation: = These Columns Get 0ed Out

Review of Linear Algebra (Cont.) SVD Reduced Representation: =

Review of Linear Algebra (Cont.) SVD Reduced Representation: = Also, Some of These May be 0

Review of Linear Algebra (Cont.) SVD Compact Representation: =

Review of Linear Algebra (Cont.) SVD Compact Representation: = These Get 0ed Out

Review of Linear Algebra (Cont.) SVD Compact Representation: =

Review of Linear Algebra (Cont.)

Recall Linear Algebra (Cont.) Moore-Penrose Generalized Inverse: For

Recall Linear Algebra (Cont.) Easy to see this satisfies the definition of Generalized (Pseudo) Inverse symmetric

Recall Linear Algebra (Cont.)

Recall Multivar. Prob.

Outer Product Representation:, Where:

Recall Multivar. Prob.

PCA as an Optimization Problem Find Direction of Greatest Variability:

PCA as an Optimization Problem Find Direction of Greatest Variability:

PCA as an Optimization Problem Find Direction of Greatest Variability: Raw Data

PCA as an Optimization Problem Find Direction of Greatest Variability: Mean Residuals (Shift to Origin)

PCA as an Optimization Problem Find Direction of Greatest Variability: Mean Residuals (Shift to Origin)

PCA as an Optimization Problem Find Direction of Greatest Variability: Centered Data

PCA as an Optimization Problem Find Direction of Greatest Variability: Centered Data Projections

PCA as an Optimization Problem Find Direction of Greatest Variability: Centered Data Projections Direction Vector

PCA as Optimization (Cont.) Find Direction of Greatest Variability: Given a Direction Vector, (i.e. ) (Variable, Over Which Will Optimize)

PCA as Optimization (Cont.) Find Direction of Greatest Variability: Given a Direction Vector, (i.e. ) Projection of in the Direction :

PCA as Optimization (Cont.) Find Direction of Greatest Variability: Given a Direction Vector, (i.e. ) Projection of in the Direction : Variability in the Direction :

PCA as Optimization (Cont.) Find Direction of Greatest Variability: Given a Direction Vector, (i.e. ) Projection of in the Direction : Variability in the Direction :

PCA as Optimization (Cont.) Find Direction of Greatest Variability: Given a Direction Vector, (i.e. ) Projection of in the Direction : Variability in the Direction :

PCA as Optimization (Cont.) Variability in the Direction :

PCA as Optimization (Cont.) Variability in the Direction : i.e. (Proportional to) a Quadratic Form in the Covariance Matrix

PCA as Optimization (Cont.) Variability in the Direction : i.e. (Proportional to) a Quadratic Form in the Covariance Matrix Simple Solution Comes from the Eigenvalue Representation of :

PCA as Optimization (Cont.) Variability in the Direction : i.e. (Proportional to) a Quadratic Form in the Covariance Matrix Simple Solution Comes from the Eigenvalue Representation of : Where is Orthonormal, &

PCA as Optimization (Cont.) Variability in the Direction :

PCA as Optimization (Cont.) Variability in the Direction : But

PCA as Optimization (Cont.) Variability in the Direction : But = “ Transform of ”

PCA as Optimization (Cont.) Variability in the Direction : But = “ Transform of ” = “ Rotated into Coordinates”,

PCA as Optimization (Cont.) Variability in the Direction : But = “ Transform of ” = “ Rotated into Coordinates”, and the Diagonalized Quadratic Form Becomes

PCA as Optimization (Cont.) Now since is an Orthonormal Basis Matrix, and

PCA as Optimization (Cont.) Now since is an Orthonormal Basis Matrix, and So the Rotation Gives a Distribution of the Energy of Over the Eigen-Directions

PCA as Optimization (Cont.) Now since is an Orthonormal Basis Matrix, and So the Rotation Gives a Distribution of the Energy of Over the Eigen-Directions And is Max’d (Over ), by Putting all Energy in the “Largest Direction”, i.e., Where “Eigenvalues are Ordered”,

PCA as Optimization (Cont.)

Iterated PCA Visualization

PCA as Optimization (Cont.)

Recall Toy Example

PCA as Optimization (Cont.) Recall Toy Example Empirical (Sample) EigenVectors

PCA as Optimization (Cont.) Recall Toy Example Theoretical Distribution

PCA as Optimization (Cont.) Recall Toy Example Theoretical Distribution & Eigenvectors

PCA as Optimization (Cont.) Recall Toy Example Empirical (Sample) EigenVectors Theoretical Distribution & Eigenvectors Different!

Connect Math to Graphics 2-d Toy Example From First Class Meeting Simple, Visualizable Descriptor Space 2-d Curves as Data In Object Space

Connect Math to Graphics 2-d Toy Example Data Points (Curves)

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example Residuals from Mean = Data - Mean

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example PC1 Projections Best 1-d Approximations of Data

Connect Math to Graphics (Cont.) 2-d Toy Example PC1 Residuals

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example PC2 Projections (= PC1 Resid’s) 2 nd Best 1-d Approximations of Data

Connect Math to Graphics (Cont.) 2-d Toy Example PC2 Residuals = PC1 Projections

Connect Math to Graphics (Cont.) Note for this 2-d Example: PC1 Residuals = PC2 Projections PC2 Residuals = PC1 Projections (i.e. colors common across these pics)

PCA Redistribution of Energy

PCA Redist ’ n of Energy (Cont.)

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

PCA Redist ’ n of Energy (Cont.) Have already studied this decomposition (recall curve e.g.)

PCA Redist ’ n of Energy (Cont.) Have already studied this decomposition (recall curve e.g.) Variation (SS) due to Mean (% of total)

PCA Redist ’ n of Energy (Cont.) Have already studied this decomposition (recall curve e.g.) Variation (SS) due to Mean (% of total) Variation (SS) of Mean Residuals (% of total)

PCA Redist ’ n of Energy (Cont.) Now Decompose SS About the Mean where: Note Inner Products this time

PCA Redist ’ n of Energy (Cont.) Now Decompose SS About the Mean where: Recall: Can Commute Matrices Inside Trace

PCA Redist ’ n of Energy (Cont.) Now Decompose SS About the Mean where: Recall: Cov Matrix is Outer Product

PCA Redist ’ n of Energy (Cont.) Now Decompose SS About the Mean where: i.e. Energy is Expressed in Trace of Cov Matrix

PCA Redist ’ n of Energy (Cont.) (Using Eigenvalue Decomp. Of Cov Matrix)

PCA Redist ’ n of Energy (Cont.) (Commute Matrices Within Trace)

PCA Redist ’ n of Energy (Cont.) (Since Basis Matrix is Orthonormal)

PCA Redist ’ n of Energy (Cont.) Eigenvalues Provide Atoms of SS Decompos’n

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

PCA Redist ’ n of Energy (Cont.) Eigenvalues Provide Atoms of SS Decompos’n

PCA Redist ’ n of Energy (Cont.) Eigenvalues Provide Atoms of SS Decompos’n Useful Plots are: Power Spectrum: vs.

PCA Redist ’ n of Energy (Cont.) Eigenvalues Provide Atoms of SS Decompos’n Useful Plots are: Power Spectrum: vs. log Power Spectrum: vs. (Very Useful When Are Orders of Mag. Apart)

PCA Redist ’ n of Energy (Cont.) Eigenvalues Provide Atoms of SS Decompos’n Useful Plots are: Power Spectrum: vs. log Power Spectrum: vs. Cumulative Power Spectrum: vs.

PCA Redist ’ n of Energy (Cont.) Eigenvalues Provide Atoms of SS Decompos’n Useful Plots are: Power Spectrum: vs. log Power Spectrum: vs. Cumulative Power Spectrum: vs. Note PCA Gives SS’s for Free (As Eigenval’s), But Watch Factors of

PCA Redist ’ n of Energy (Cont.) Note, have already considered some of these Useful Plots:

PCA Redist ’ n of Energy (Cont.) Note, have already considered some of these Useful Plots: Power Spectrum

PCA Redist ’ n of Energy (Cont.) Note, have already considered some of these Useful Plots: Power Spectrum Cumulative Power Spectrum

PCA Redist ’ n of Energy (Cont.) Note, have already considered some of these Useful Plots: Power Spectrum Cumulative Power Spectrum Common Terminology: Power Spectrum is Called “Scree Plot” Kruskal (1964) Cattell (1966) (all but name “scree”) (1 st Appearance of name???)

Participant Presentation Anna Zhao Surviving in the NBA