Spatial autoregressive methods Nr245 Austin Troy Based on Spatial Analysis by Fortin and Dale, Chapter 5.

Slides:



Advertisements
Similar presentations
Spatial autoregressive methods
Advertisements

Managerial Economics in a Global Economy
The Multiple Regression Model.
Objectives 10.1 Simple linear regression
Hypothesis Testing Steps in Hypothesis Testing:
1 Regression Models & Loss Reserve Variability Prakash Narayan Ph.D., ACAS 2001 Casualty Loss Reserve Seminar.
4.3 Confidence Intervals -Using our CLM assumptions, we can construct CONFIDENCE INTERVALS or CONFIDENCE INTERVAL ESTIMATES of the form: -Given a significance.
Basic geostatistics Austin Troy.
Spatial Autocorrelation Basics NR 245 Austin Troy University of Vermont.
Dealing with Spatial Autocorrelation
Evaluation (practice). 2 Predicting performance  Assume the estimated error rate is 25%. How close is this to the true error rate?  Depends on the amount.
Correlation and Autocorrelation
PSY 307 – Statistics for the Behavioral Sciences
BCOR 1020 Business Statistics Lecture 17 – March 18, 2008.
Chapter 12 Simple Regression
Simulation Modeling and Analysis Session 12 Comparing Alternative System Designs.
SA basics Lack of independence for nearby obs
Basics of regression analysis
Today Concepts underlying inferential statistics
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
Applications in GIS (Kriging Interpolation)
AAEC 4302 ADVANCED STATISTICAL METHODS IN AGRICULTURAL RESEARCH Chapter 13.3 Multicollinearity.
Lorelei Howard and Nick Wright MfD 2008
Method of Soil Analysis 1. 5 Geostatistics Introduction 1. 5
Lecture 5 Correlation and Regression
Structural Equation Modeling Continued: Lecture 2 Psy 524 Ainsworth.
Hypothesis Testing and T-Tests. Hypothesis Tests Related to Differences Copyright © 2009 Pearson Education, Inc. Chapter Tests of Differences One.
Multiple Linear Regression Response Variable: Y Explanatory Variables: X 1,...,X k Model (Extension of Simple Regression): E(Y) =  +  1 X 1 +  +  k.
Physics 114: Lecture 15 Probability Tests & Linear Fitting Dale E. Gary NJIT Physics Department.
Regression Method.
STATISTICS: BASICS Aswath Damodaran 1. 2 The role of statistics Aswath Damodaran 2  When you are given lots of data, and especially when that data is.
CORRELATION & REGRESSION
Sampling Distribution of the Mean Central Limit Theorem Given population with and the sampling distribution will have: A mean A variance Standard Error.
Some matrix stuff.
One Sample Inf-1 If sample came from a normal distribution, t has a t-distribution with n-1 degrees of freedom. 1)Symmetric about 0. 2)Looks like a standard.
Geo479/579: Geostatistics Ch12. Ordinary Kriging (1)
Basic geostatistics Austin Troy.
Inferential Statistics 2 Maarten Buis January 11, 2006.
1 Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved. Section 7.4 Estimation of a Population Mean  is unknown  This section presents.
From Theory to Practice: Inference about a Population Mean, Two Sample T Tests, Inference about a Population Proportion Chapters etc.
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Copyright © 2005 by the McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Managerial Economics Thomas Maurice eighth edition Chapter 4.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 8-1 Confidence Interval Estimation.
CHAPTER 5 SIGNAL SPACE ANALYSIS
Chapter 10 The t Test for Two Independent Samples
Lecture 6: Point Interpolation
Trees Example More than one variable. The residual plot suggests that the linear model is satisfactory. The R squared value seems quite low though,
June 30, 2008Stat Lecture 16 - Regression1 Inference for relationships between variables Statistics Lecture 16.
Estimating Volatilities and Correlations
University of Colorado Boulder ASEN 5070 Statistical Orbit determination I Fall 2012 Professor George H. Born Professor Jeffrey S. Parker Lecture 9: Least.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 8-1 Chapter 8 Confidence Interval Estimation Business Statistics: A First Course 5 th Edition.
Simple and multiple regression analysis in matrix form Least square Beta estimation Beta Simple linear regression Multiple regression with two predictors.
6-1 Copyright © 2014, 2011, and 2008 Pearson Education, Inc.
Chapter 8 Confidence Interval Estimation Statistics For Managers 5 th Edition.
Université d’Ottawa / University of Ottawa 2003 Bio 8102A Applied Multivariate Biostatistics L4.1 Lecture 4: Multivariate distance measures l The concept.
Confidence Intervals.
Spatial statistics: Spatial Autocorrelation
Chapter 6 Inferences Based on a Single Sample: Estimation with Confidence Intervals Slides for Optional Sections Section 7.5 Finite Population Correction.
Basic Estimation Techniques
Chapter 5 Part B: Spatial Autocorrelation and regression modelling.
Fundamentals of regression analysis
Chapter 11 Simple Regression
Chapter 13 Simple Linear Regression
Basic Estimation Techniques
Chapter 9 Hypothesis Testing.
Statistical Inference about Regression
Presentation transcript:

Spatial autoregressive methods Nr245 Austin Troy Based on Spatial Analysis by Fortin and Dale, Chapter 5

Autcorrelation types None: independence Spatial independence, functional dependence True autocorrelation>> inherent autoregressive Functional autocorr>> induced autoregressive

Autocorrelation types Double autoregressive Notice there are now two autocorrelation parameters  - x and  -z

Effects? Standard test statistics become “too liberal”—more significant results than the data justify Because observations are not totally independent have lower actual degrees of freedom, or lower “effective sample size”: n’ instead of n; since t stat denominator = s/n, if n is too big it inflates the t statistic

What to do? Non-effective Why not just adjust up the significance level? E.g. 99% instead of 95%? Because don’t how by how much to adjust without further information. Could end up with a test that is way too conservative Why not just adjust sampling to only include “independent samples?” Because wasteful of data and because easy to mistake “critical distance to independence”

Best approach: Adjust effective sample size For large sample sizes –So for instance n=1000 and ro=.4 means n’=429 Problem is that, to be useful, autoregressive model (ro parameter) has to be an effective descriptor of the structure of autocorrelation of the data

Moving average models How calculated depends on “order” A simple model for adjusting sample size: first order autoregressive model, only immediate (first order) neighbors are correlated with ro>0. All other pairs are zero. In such a model x i is a function of x i+1 and x i-1 Hence half the info for x i is in each neighbor; produce ro=.5 for large n and n’=n/2. An n order model can take form Translates into generalized matrix form With variance covariance matrix

Moving average When you increase the order, calculating sample size gets complicated; e.g. second order model, where two ro parameters now Important point: If there are several different levels of autocorrelation (  k ), each  k must be incorporated even if non-significant; this can have a huge impact on the calculation of effective sample size Fortin and Dale recommend not using moving average approach because very sensitive to irregularities in the data and can produce a wide range of estimates

Two dimensional approaches Problem with MA approach as it was just presented is assumes one-dimensionality In spatial data, xi depends on all neighbors most likely Two best ways for dealing with this: –Simultaneous autoregressive models (SAR) –Conditional autoregressive models (CAR)

SAR Based on concept of set of simultaneous equations to be solved. In this x i and x i-1 are each defined by their own equations Where x is a vector and is linearly dependent on a vector of underlying variables z 1, z 2 z 3 …. Given as matrix Z, u is a vector non-independent error terms with mean zero and var-covar matrix C Spatial autocorrelation enters via u where Here e is independent error term and W is neighbor weights standardized to row totals of 1. W is not necessarily symmetrical, allowing for inclusion of anisotropy. W ij is >0 if values at location i is not independent of value at location j

SAR This yields the model With variance covariance matrix (from u) Note how similar to MA—difference is no inverse in formula The elements of C are variances From Fortin and Dale p. 231

CAR More commonly used in spatial statistics Not based on spatial dependence per se; instead probability of a certain value is conditional on neighbor values Similar to SAR, but requires that weight matrix be symmetrical Here Where  is the autocorrelation parameter and V is a symmetrical weight matrix Any SAR process is a CAR process if V= W + W T – W T W