Lecture 17 Factor Analysis. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture 03Probability and.

Slides:



Advertisements
Similar presentations
Lecture 1 Describing Inverse Problems. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture 03Probability.
Advertisements

Lecture 20 Continuous Problems Linear Operators and Their Adjoints.
Lecture 10 Nonuniqueness and Localized Averages. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Environmental Data Analysis with MatLab Lecture 15: Factor Analysis.
Lecture 14 Nonlinear Problems Grid Search and Monte Carlo Methods.
General Linear Model With correlated error terms  =  2 V ≠  2 I.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Lecture 13 L1 , L∞ Norm Problems and Linear Programming
Lecture 23 Exemplary Inverse Problems including Earthquake Location.
Environmental Data Analysis with MatLab
Lecture 22 Exemplary Inverse Problems including Filter Design.
Lecture 18 Varimax Factors and Empircal Orthogonal Functions.
Environmental Data Analysis with MatLab Lecture 16: Orthogonal Functions.
Least Squares example There are 3 mountains u,y,z that from one site have been measured as 2474 ft., 3882 ft., and 4834 ft.. But from u, y looks 1422 ft.
Lecture 7 Backus-Gilbert Generalized Inverse and the Trade Off of Resolution and Variance.
Lecture 3 Probability and Measurement Error, Part 2.
Lecture 2 Probability and Measurement Error, Part 1.
Lecture 5 A Priori Information and Weighted Least Squared.
Environmental Data Analysis with MatLab Lecture 17: Covariance and Autocorrelation.
Lecture 19 Continuous Problems: Backus-Gilbert Theory and Radon’s Problem.
Lecture 15 Nonlinear Problems Newton’s Method. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 4 The L 2 Norm and Simple Least Squares. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 16 Nonlinear Problems: Simulated Annealing and Bootstrap Confidence Intervals.
Lecture 9 Inexact Theories. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture 03Probability and.
Lecture 21 Continuous Problems Fréchet Derivatives.
Lecture 24 Exemplary Inverse Problems including Vibrational Problems.
Lecture 6 Resolution and Generalized Inverses. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Some useful linear algebra. Linearly independent vectors span(V): span of vector space V is all linear combinations of vectors v i, i.e.
Environmental Data Analysis with MatLab Lecture 5: Linear Models.
The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y: Rank (X Y) =min (rank (X), rank (Y)) A = C S.
Lecture 12 Equality and Inequality Constraints. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Math for CSLecture 41 Linear Least Squares Problem Over-determined systems Minimization problem: Least squares norm Normal Equations Singular Value Decomposition.
1 Algorithms for Large Data Sets Ziv Bar-Yossef Lecture 4 March 30, 2005
Lecture 8 The Principle of Maximum Likelihood. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Environmental Data Analysis with MatLab Lecture 24: Confidence Limits of Spectra; Bootstraps.
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Clustering In Large Graphs And Matrices Petros Drineas, Alan Frieze, Ravi Kannan, Santosh Vempala, V. Vinay Presented by Eric Anderson.
Lecture 11 Vector Spaces and Singular Value Decomposition.
Multimedia Databases LSI and SVD. Text - Detailed outline text problem full text scanning inversion signature files clustering information filtering and.
Lecture 20 Empirical Orthogonal Functions and Factor Analysis.
Linear Least Squares QR Factorization. Systems of linear equations Problem to solve: M x = b Given M x = b : Is there a solution? Is the solution unique?
6 6.3 © 2012 Pearson Education, Inc. Orthogonality and Least Squares ORTHOGONAL PROJECTIONS.
Linear and generalised linear models
Basics of regression analysis
Today Wrap up of probability Vectors, Matrices. Calculus
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka Virginia de Sa (UCSD) Cogsci 108F Linear.
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
Summarized by Soo-Jin Kim
AN ORTHOGONAL PROJECTION
SVD: Singular Value Decomposition
تهیه کننده : نرگس مرعشی استاد راهنما : جناب آقای دکتر جمشید شنبه زاده.
Lecture 16 - Approximation Methods CVEN 302 July 15, 2002.
Geology 5670/6670 Inverse Theory 20 Feb 2015 © A.R. Lowry 2015 Read for Mon 23 Feb: Menke Ch 9 ( ) Last time: Nonlinear Inversion Solution appraisal.
Instructor: Mircea Nicolescu Lecture 8 CS 485 / 685 Computer Vision.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Singular Value Decomposition and Numerical Rank. The SVD was established for real square matrices in the 1870’s by Beltrami & Jordan for complex square.
Environmental Data Analysis with MatLab 2 nd Edition Lecture 22: Linear Approximations and Non Linear Least Squares.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
Lecture 8:Eigenfaces and Shared Features
Singular Value Decomposition
Unfolding Problem: A Machine Learning Approach
Some useful linear algebra
~ Least Squares example
Outline Singular Value Decomposition Example of PCA: Eigenfaces.
~ Least Squares example
Dynamic modeling of gene expression data
Unfolding with system identification
Presentation transcript:

Lecture 17 Factor Analysis

Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture 03Probability and Measurement Error, Part 2 Lecture 04The L 2 Norm and Simple Least Squares Lecture 05A Priori Information and Weighted Least Squared Lecture 06Resolution and Generalized Inverses Lecture 07Backus-Gilbert Inverse and the Trade Off of Resolution and Variance Lecture 08The Principle of Maximum Likelihood Lecture 09Inexact Theories Lecture 10Nonuniqueness and Localized Averages Lecture 11Vector Spaces and Singular Value Decomposition Lecture 12Equality and Inequality Constraints Lecture 13L 1, L ∞ Norm Problems and Linear Programming Lecture 14Nonlinear Problems: Grid and Monte Carlo Searches Lecture 15Nonlinear Problems: Newton’s Method Lecture 16Nonlinear Problems: Simulated Annealing and Bootstrap Confidence Intervals Lecture 17Factor Analysis Lecture 18Varimax Factors, Empircal Orthogonal Functions Lecture 19Backus-Gilbert Theory for Continuous Problems; Radon’s Problem Lecture 20Linear Operators and Their Adjoints Lecture 21Fréchet Derivatives Lecture 22 Exemplary Inverse Problems, incl. Filter Design Lecture 23 Exemplary Inverse Problems, incl. Earthquake Location Lecture 24 Exemplary Inverse Problems, incl. Vibrational Problems

Purpose of the Lecture Introduce Factor Analysis Work through an example

Part 1 Factor Analysis

source A ocean sediment source B s4s4 s2s2 s3s3 s1s1

sample matrix S S arranged row-wise but we’ll use a column vector s (i) for individual samples)

theory samples are a linear mixture of sources S = C F

theory samples are a linear mixture of sources S = C F samples contain “elements”

theory samples are a linear mixture of sources S = C F sources called “factors” factors contain “elements”

factor matrix F F arranged row-wise but we’ll use a column vector f (i) for individual factors

theory samples are a linear mixture of sources S = C F coefficients called “loadings”

loading matrix C

inverse problem given S find C and F so that S=CF

very non-unique given T with inverse T -1 if S=CF then S=[C T -1 ][TF] =C’F’

very non-unique so a priori information needed to select a solution

simplicity what is the minimum number of factors needed call that number p

does S span the full space of M elements? or just a p –dimensional subspace?

E3E3 E2E2 s1s1 s3s3 A B s1s1 s4s4 E1E1

we know how to answer this question p is the number of non-zero singular values

E3E3 E2E2 E1E1 s2s2 s3s3 A B s1s1 s4s4 v2v2 v1v1 v3v3

SVD identifies a subspace but the SVD factors f (i) = v (i) i=1, p not unique usually not the “best”

factor f (1) v with the largest singular value usually near the mean sample sample mean minimize eigenvector minimize

factor f (1) v with the largest singular value usually near the mean sample sample mean minimize eigenvector minimize about the same if samples are clustered

s3s3 f (1) s1s1 s4s4 f (2) s4s4 s3s3 s2s2 s2s2 s1s1 E3E3 E2E2 E1E1 E3E3 E2E2 E1E1 (A) (B)

[U, LAMBDA, V] = svd(S,0); lambda = diag(LAMBDA); F = V'; C = U*LAMBDA; in MatLab

[U, LAMBDA, V] = svd(S,0); lambda = diag(LAMBDA); F = V'; C = U*LAMBDA; “economy” calculation LAMBDA is M⨉M in MatLab

since samples have measurement noise probably no exactly singular values just very small ones so pick p for which S≈CF is an adequate approximation

Atlantic Rock Dataset (several thousand more rows) SiO 2 TiO 2 Al 2 O 3 FeO t MgO CaO Na 2 O K 2 O

Al Ti0 2 Al Si0 2 K20K20 Fe0 Mg0 Al A)B) C)D)

singular values, λ i index, i

f (5) f (2) f (3) f (4) SiO 2 TiO 2 Al 2 O 3 FeO total MgO CaO Na 2 O K2OK2O

C2C2 C3C3 C4C4