A Data Partitioning Scheme for Spatial Regression

Slides:



Advertisements
Similar presentations
Detecting Statistical Interactions with Additive Groves of Trees
Advertisements

Sampling Design, Spatial Allocation, and Proposed Analyses Don Stevens Department of Statistics Oregon State University.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Random Forest Predrag Radenković 3237/10
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Multiple Regression Analysis
Methods for Improving Protein Disorder Prediction Slobodan Vucetic1, Predrag Radivojac3, Zoran Obradovic3, Celeste J. Brown2, Keith Dunker2 1 School of.
Regression Chapter 10 Understandable Statistics Ninth Edition By Brase and Brase Prepared by Yixun Shi Bloomsburg University of Pennsylvania.
© Prentice Hall1 DATA MINING Introductory and Advanced Topics Part II Margaret H. Dunham Department of Computer Science and Engineering Southern Methodist.
Quantitative Business Analysis for Decision Making Simple Linear Regression.
General Mining Issues a.j.m.m. (ton) weijters Overfitting Noise and Overfitting Quality of mined models (some figures are based on the ML-introduction.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Part I: Classification and Bayesian Learning
Applications in GIS (Kriging Interpolation)
Ensemble Learning (2), Tree and Forest
1 Comparison of Discrimination Methods for the Classification of Tumors Using Gene Expression Data Presented by: Tun-Hsiang Yang.
Radial Basis Function Networks
TIME SERIES by H.V.S. DE SILVA DEPARTMENT OF MATHEMATICS
Experimental Design An Experimental Design is a plan for the assignment of the treatments to the plots in the experiment Designs differ primarily in the.
$88.65 $ $22.05/A profit increase Improving Wheat Profits Eakly, OK Irrigated, Behind Cotton.
Chapter 11 – Neural Networks COMP 540 4/17/2007 Derek Singer.
Why Is It There? Getting Started with Geographic Information Systems Chapter 6.
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
Correlation & Regression
Gap filling of eddy fluxes with artificial neural networks
Regression Regression relationship = trend + scatter
A comparison of the ability of artificial neural network and polynomial fitting was carried out in order to model the horizontal deformation field. It.
Control of Experimental Error Blocking - –A block is a group of homogeneous experimental units –Maximize the variation among blocks in order to minimize.
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
Ivica Dimitrovski 1, Dragi Kocev 2, Suzana Loskovska 1, Sašo Džeroski 2 1 Faculty of Electrical Engineering and Information Technologies, Department of.
Speech Lab, ECE, State University of New York at Binghamton  Classification accuracies of neural network (left) and MXL (right) classifiers with various.
Classification Ensemble Methods 1
How Good is a Model? How much information does AIC give us? –Model 1: 3124 –Model 2: 2932 –Model 3: 2968 –Model 4: 3204 –Model 5: 5436.
1 Introduction to Predictive Learning Electrical and Computer Engineering LECTURE SET 8 Combining Methods and Ensemble Learning.
Computational Intelligence: Methods and Applications Lecture 15 Model selection and tradeoffs. Włodzisław Duch Dept. of Informatics, UMK Google: W Duch.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
Introduction Many problems in Engineering, Management, Health Sciences and other Sciences involve exploring the relationships between two or more variables.
Chapter 11 – Neural Nets © Galit Shmueli and Peter Bruce 2010 Data Mining for Business Intelligence Shmueli, Patel & Bruce.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Data Transformation: Normalization
MATH-138 Elementary Statistics
CEE 6410 Water Resources Systems Analysis
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Neural Networks Winter-Spring 2014
第 3 章 神经网络.
Chapter 11: Simple Linear Regression
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
CSE 4705 Artificial Intelligence
PCB 3043L - General Ecology Data Analysis.
Prediction of Coal Free-Swelling Index using Mathematical Modelling
A Simple Artificial Neuron
Precision Nutrient Management: Grid-Sampling Basis
Machine Learning Today: Reading: Maria Florina Balcan
Regression Analysis Week 4.
Hyperparameters, bias-variance tradeoff, validation
Efficient Distribution-based Feature Search in Multi-field Datasets Ohio State University (Shen) Problem: How to efficiently search for distribution-based.
Chap. 7 Regularization for Deep Learning (7.8~7.12 )
Spatial interpolation
Linear Model Selection and regularization
What are BLUP? and why they are useful?
Introduction to Radial Basis Function Networks
Model generalization Brief summary of methods
Neural networks (1) Traditional multi-layer perceptrons
Jozsef Csicsvari, Hajime Hirase, Akira Mamiya, György Buzsáki  Neuron 
Primer on Neural networks
Advisor: Dr.vahidipour Zahra salimian Shaghayegh jalali Dec 2017
Creating Trends for Reservoir Modelling Using ANN
Presentation transcript:

A Data Partitioning Scheme for Spatial Regression S. Vucetic1, T. Fiez2 and Z. Obradovic1, 1 School of Electrical Engineering and Computer Science 2 Department of Crop and Soil Sciences Washington State University, Pullman

Purpose First step in training is to partition available data randomly into training, validation and test subsets Spatial data are a collection of variables whose dependence is strongly tied to a spatial location observations close to each other are more likely to be similar than observations widely separated in space. Training (e.g. using neural networks) on spatial data is likely to produce spatially correlated errors.

Goal Large amounts of agricultural spatial data are gathered using global positioning systems. Objective - To predict wheat yield from spatial attributes in order to extrapolate this knowledge to different agricultural sites, or to the same site but in different years. Goal - To design an appropriate procedure for training neural networks and to properly estimate their extrapolation properties

Method Choice of test subset To decrease the influence of spatial proximity and estimate the predictor accuracy, the test subset should be spatially separated from the training data WEST EAST

Method Use of a randomly selected validation subset will result in overfitting The training portion of the field is partitioned into squares of size MM Half of these squares are randomly assigned for use in training and the rest for validation

BLACK – training subset GRAY – validation subset Figure 2. Spatial partitioning of east half of wheat field into squares of size M=100m. Black and gray squares represent training and validation data while white represents missing data

Method Choice of block size, M M should be sufficiently large to minimize the influence of spatial correlation between training and validation data, M should be small enough to provide a training set representative of the whole field. Solution Select M to be within a range where correlograms (plot of the correlation coefficient as a function of the separation distance between data points) of all features approach zero.

Method Spatial bagging Train a number of neural networks for different random assignments of squares into training and validation subsets Final prediction is average prediction from all constructed neural networks.

Figure 3. Histograms of 8 topographic features Data Set Data set - A precision agriculture database containing a 10 x 10 m grid of 8 topographic attributes and winter wheat yield totaling 24,592 data points from a 220 ha field located near Pullman, WA Figure 3. Histograms of 8 topographic features Figure 4. Correlograms for variables whose histograms are shown in Figure 3 Distance [m]

Results Comparison of Random Testing (RTS) vs Spatially Disjoint Testing (SDT) MODEL MSE   Train RTS SDTS Mean predictor 369 366 390 Linear 322 317 371 NN 5 hidden nodes 253 267 370 NN 15 hidden nodes 218 238 375 Random testing is missleading

Results Influence of block size on prediction MODEL MSE EPOCHS Mean predictor 390 - Linear 370  4 NN 5 hidden nodes M = 10m M = 40m M = 100m M = 200m 37114 14731 36517 4319 36515 3014 36421 2510 M determined by correlogram range, M=[40, 200] m, resulted in better MSE and shorter training time

Results Comparison between bagging and spatial bagging METHOD MSE NN 5 hidden nodes NN 15 hidden Bagging 357 351 Spatial Bagging M = 40m M = 100m M = 200m 344 348 342 339 Spatial bagging results in better prediction The margin between the best models and a trivial predictor was fairly small

Conclusions Spatial partitioning into training and validation subsets can lead to substantial improvements in learning speed, and increased predictability, as compared to the traditional random partitioning. Correlograms can be used to determine the parameters of such spatial data partitioning. Spatial bagging can improve generalization capabilities as compared to bagging.