Meet the professor Friday, January 23 at SFU 4:30 Beer and snacks reception.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Pattern Recognition and Machine Learning
Spatial point patterns and Geostatistics an introduction
Spatial point patterns and Geostatistics an introduction
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Periodograms Bartlett Windows Data Windowing Blackman-Tukey Resources:
Characterizing Non- Gaussianities or How to tell a Dog from an Elephant Jesús Pando DePaul University.
The General Linear Model. The Simple Linear Model Linear Regression.
Basic geostatistics Austin Troy.
Spectral analysis for point processes. Error bars. Bijan Pesaran Center for Neural Science New York University.
Spatial prediction Stat 498A, Winter The geostatistical model Gaussian process  (s)=EZ(s) Var Z(s) < ∞ Z is strictly stationary if Z is weakly.
Nonstationary covariance structures II NRCSE. Drawbacks with deformation approach Oversmoothing of large areas Local deformations not local part of global.
SUMS OF RANDOM VARIABLES Changfei Chen. Sums of Random Variables Let be a sequence of random variables, and let be their sum:
Spatial Statistics IV Stat 518 Sp08. Recall Method of moments: square of all pairwise differences, smoothed over lag bins Problems: Not necessarily a.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
The Fourier transform. Properties of Fourier transforms Convolution Scaling Translation.
Applied Geostatistics
Parametric Inference.
Continuity and covariance Recall that for stationary processes so if C is continuous then Z is mean square continuous. To get results about sample paths.
Variograms/Covariances and their estimation
Tch-prob1 Chapter 4. Multiple Random Variables Ex Select a student’s name from an urn. S In some random experiments, a number of different quantities.
Deterministic Solutions Geostatistical Solutions
7. Least squares 7.1 Method of least squares K. Desch – Statistical methods of data analysis SS10 Another important method to estimate parameters Connection.
Nonstationary covariance structures II NRCSE. Drawbacks with deformation approach Oversmoothing of large areas Local deformations not local part of global.
Global processes Problems such as global warming require modeling of processes that take place on the globe (an oriented sphere). Optimal prediction of.
Space-time processes NRCSE. Separability Separable covariance structure: Cov(Z(x,t),Z(y,s))=C S (x,y)C T (s,t) Nonseparable alternatives Temporally varying.
Bayesian kriging Instead of estimating the parameters, we put a prior distribution on them, and update the distribution using the data. Model: Prior: Posterior:
STAT 592A(UW) 526 (UBC-V) 890-4(SFU) Spatial Statistical Methods NRCSE.
Computer vision: models, learning and inference
Spatial Statistics III Stat 518 Sp08. Bayesian kriging Instead of estimating the parameters, we put a prior distribution on them, and update the distribution.
Statistical Tools for Environmental Problems NRCSE.
Introduction To Signal Processing & Data Analysis
Lecture II-2: Probability Review
Review of Probability.
Goals For This Class Quickly review of the main results from last class Convolution and Cross-correlation Discrete Fourier Analysis: Important Considerations.
Random Sampling, Point Estimation and Maximum Likelihood.
Geo479/579: Geostatistics Ch12. Ordinary Kriging (1)
Time Series Spectral Representation Z(t) = {Z 1, Z 2, Z 3, … Z n } Any mathematical function has a representation in terms of sin and cos functions.
Mathematical Preliminaries. 37 Matrix Theory Vectors nth element of vector u : u(n) Matrix mth row and nth column of A : a(m,n) column vector.
Geographic Information Science
Estimation of the spectral density function. The spectral density function, f( ) The spectral density function, f(x), is a symmetric function defined.
Semivariogram Analysis and Estimation Tanya, Nick Caroline.
Geology 5670/6670 Inverse Theory 21 Jan 2015 © A.R. Lowry 2015 Read for Fri 23 Jan: Menke Ch 3 (39-68) Last time: Ordinary Least Squares Inversion Ordinary.
Estimation of the spectral density function. The spectral density function, f( ) The spectral density function, f(x), is a symmetric function defined.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
Brief Review Probability and Statistics. Probability distributions Continuous distributions.
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
Short course on space-time modeling Instructors: Peter Guttorp Johan Lindström Paul Sampson.
Stochastic Hydrology Random Field Simulation Professor Ke-Sheng Cheng Department of Bioenvironmental Systems Engineering National Taiwan University.
Geostatistics GLY 560: GIS for Earth Scientists. 2/22/2016UB Geology GLY560: GIS Introduction Premise: One cannot obtain error-free estimates of unknowns.
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and  2 Now, we need procedures to calculate  and  2, themselves.
Week 21 Order Statistics The order statistics of a set of random variables X 1, X 2,…, X n are the same random variables arranged in increasing order.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Space-time processes NRCSE. Separability Separable covariance structure: Cov(Z(x,t),Z(y,s))=C S (x,y)C T (s,t) Nonseparable alternatives Temporally varying.
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
CS Statistical Machine learning Lecture 7 Yuan (Alan) Qi Purdue CS Sept Acknowledgement: Sargur Srihari’s slides.
Biointelligence Laboratory, Seoul National University
Time Series Spectral Representation
Probability Theory and Parameter Estimation I
Estimation of the spectral density function
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
NRCSE 2. Covariances.
Nonstationary covariance structures II
Paul D. Sampson Peter Guttorp
Stochastic Hydrology Random Field Simulation
The Empirical FT. What is the large sample distribution of the EFT?
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Probabilistic Surrogate Models
Presentation transcript:

Meet the professor Friday, January 23 at SFU 4:30 Beer and snacks reception

Spatial Covariance NRCSE

Valid covariance functions Bochner’s theorem: The class of covariance functions is the class of positive definite functions C: Why?

By the spectral representation any isotropic continuous correlation on R d is of the form By isotropy, the expectation depends only on the distribution G of. Let Y be uniform on the unit sphere. Then Spectral representation

Isotropic correlation J v (u) is a Bessel function of the first kind and order v. Hence and in the case d=2 (Hankel transform)

The Bessel function J 0 –0.403

The exponential correlation A commonly used correlation function is  (v) = e –v/ . Corresponds to a Gaussian process with continuous but not differentiable sample paths. More generally,  (v) = c(v=0) + (1-c)e –v/  has a nugget c, corresponding to measurement error and spatial correlation at small distances. All isotropic correlations are a mixture of a nugget and a continuous isotropic correlation.

The squared exponential Usingyields corresponding to an underlying Gaussian field with analytic paths. This is sometimes called the Gaussian covariance, for no really good reason. A generalization is the power(ed) exponential correlation function,

The spherical Corresponding variogram nugget sillrange

The Matérn class where is a modified Bessel function of the third kind and order . It corresponds to a spatial field with [  –1] continuous derivatives  =1/2 is exponential;  =1 is Whittle’s spatial correlation; yields squared exponential.

Some other covariance/variogram families NameCovarianceVariogram Wave Rational quadratic LinearNone Power lawNone

Recall Method of moments: square of all pairwise differences, smoothed over lag bins Problems: Not necessarily a valid variogram Not very robust Estimation of variograms

A robust empirical variogram estimator (Z(x)-Z(y)) 2 is chi-squared for Gaussian data Fourth root is variance stabilizing Cressie and Hawkins:

Least squares Minimize Alternatives: fourth root transformation weighting by 1/  2 generalized least squares

Maximum likelihood Z~N n ( ,  )  =  [  (s i -s j ;  )] =  V(  ) Maximize and  maximizes the profile likelihood

A peculiar ml fit

Some more fits

All together now...

Asymptotics Increasing domain asymptotics: let region of interest grow. Station density stays the same Bad estimation at short distances, but effectively independent blocks far apart Infill asymptotics: let station density grow, keeping region fixed. Good estimates at short distances. No effectively independent blocks, so technically trickier

Stein’s result Covariance functions C 0 and C 1 are compatible if their Gaussian measures are mutually absolutely continuous. Sample at {s i, i=1,...,n}, predict at s (limit point of sampling points). Let e i (n) be kriging prediction error at s for C i, and V 0 the variance under C 0 of some random variable. If lim n V 0 (e 0 (n))=0, then

The Fourier transform

Properties of Fourier transforms Convolution Scaling Translation

Parceval’s theorem Relates space integration to frequency integration. Decomposes variability.

Aliasing Observe field at lattice of spacing . Since the frequencies  and  ’=  +2  m/  are aliases of each other, and indistinguishable. The highest distinguishable frequency is , the Nyquist frequency.

Illustration of aliasing Aliasing applet

Spectral representation Stationary processes Spectral process Y has stationary increments If F has a density f, it is called the spectral density.

Estimating the spectrum For process observed on nxn grid, estimate spectrum by periodogram Equivalent to DFT of sample covariance

Properties of the periodogram Periodogram values at Fourier frequencies (j,k)  are uncorrelated asymptotically unbiased not consistent To get a consistent estimate of the spectrum, smooth over nearby frequencies

Some common isotropic spectra Squared exponential Matérn

A simulated process

Thetford canopy heights 39-year thinned commercial plantation of Scots pine in Thetford Forest, UK Density 1000 trees/ha 36m x 120m area surveyed for crown height Focus on 32 x 32 subset

Spectrum of canopy heights

Whittle likelihood Approximation to Gaussian likelihood using periodogram: where the sum is over Fourier frequencies, avoiding 0, and f is the spectral density Takes O(N logN) operations to calculate instead of O(N 3 ).

Using non-gridded data Consider where Then Y is stationary with spectral density Viewing Y as a lattice process, it has spectral density

Estimation Let where J x is the grid square with center x and n x is the number of sites in the square. Define the tapered periodogram where. The Whittle likelihood is approximately

A simulated example

Estimated variogram