A comparison of the ability of artificial neural network and polynomial fitting was carried out in order to model the horizontal deformation field. It.

Slides:



Advertisements
Similar presentations
FUNCTION FITTING Student’s name: Ruba Eyal Salman Supervisor:
Advertisements

Artificial Intelligence 13. Multi-Layer ANNs Course V231 Department of Computing Imperial College © Simon Colton.
Ludovico Biagi & Athanasios Dermanis Politecnico di Milano, DIIAR Aristotle University of Thessaloniki, Department of Geodesy and Surveying Crustal Deformation.
On the alternative approaches to ITRF formulation. A theoretical comparison. Department of Geodesy and Surveying Aristotle University of Thessaloniki Athanasios.
CMPUT 466/551 Principal Source: CMU
Chapter 19 Confidence Intervals for Proportions.
The loss function, the normal equation,
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
OLS REGRESSION VS. NEURAL NETWORKS VS. MARS A COMPARISON R. J. Lievano E. Kyper University of Minnesota Duluth.
Radial Basis Functions
Chapter 5 NEURAL NETWORKS
Curve-Fitting Regression
Designing the configuration of the Geodetic-Geodynamic Network in Israel Gilad Even-Tzur Department of Mapping and Geo-Information Engineering Faculty.
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
The Calibration Process
13/06/13 H. Rho Slide 1 Geodetic Research Laboratory Department of Geodesy and Geomatics Engineering University of New Brunswick Evaluation of Precise.
Chapter 6 Random Error The Nature of Random Errors
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Application of CAS to geodesy: a ‘live’ approach P. Zaletnyik 1, B. Paláncz 2, J.L. Awange 3, E.W. Grafarend 4 1,2 Budapest University of Technology and.
 C. C. Hung, H. Ijaz, E. Jung, and B.-C. Kuo # School of Computing and Software Engineering Southern Polytechnic State University, Marietta, Georgia USA.
Chapter 15 Modeling of Data. Statistics of Data Mean (or average): Variance: Median: a value x j such that half of the data are bigger than it, and half.
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
A Shaft Sensorless Control for PMSM Using Direct Neural Network Adaptive Observer Authors: Guo Qingding Luo Ruifu Wang Limei IEEE IECON 22 nd International.
FIG Working Week- Bridging the Gap Between Cultures Marrakech, Morocco May Using of both fast Static and RTK Modes for GNSS.
11 CSE 4705 Artificial Intelligence Jinbo Bi Department of Computer Science & Engineering
Outline 1-D regression Least-squares Regression Non-iterative Least-squares Regression Basis Functions Overfitting Validation 2.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 9: Ways of speeding up the learning and preventing overfitting Geoffrey Hinton.
Blue – comp red - ext. blue – comp red - ext blue – comp red - ext.
SNARF: Theory and Practice, and Implications Thomas Herring Department of Earth Atmospheric and Planetary Sciences, MIT
Handwritten Recognition with Neural Network Chatklaw Jareanpon, Olarik Surinta Mahasarakham University.
Artificial Intelligence Chapter 3 Neural Networks Artificial Intelligence Chapter 3 Neural Networks Biointelligence Lab School of Computer Sci. & Eng.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION ASEN 5070 LECTURE 11 9/16,18/09.
A new Ad Hoc Positioning System 컴퓨터 공학과 오영준.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Jayne Bormann and Bill Hammond sent two velocity fields on a uniform grid constructed from their test exercise using CMM4. Hammond ’ s code.
CROSS-VALIDATION AND MODEL SELECTION Many Slides are from: Dr. Thomas Jensen -Expedia.com and Prof. Olga Veksler - CS Learning and Computer Vision.
CLIC survey and alignment 1 CLIC CES meeting STARTSLIDE CLIC SURVEY AND ALIGNMENT Hélène MAINAUD DURAND.
Neural Nets: Something you can use and something to think about Cris Koutsougeras What are Neural Nets What are they good for Pointers to some models and.
Application of a North America reference frame to the Pacific Northwest Geodetic Array (PANGA) M M Miller, V M Santillan, Geodesy Laboratory, Central Washington.
Technical Specifications for Monitoring Ground Displacements at a National Highway Project K. Lakakis, P. Savvaidis and I. Ifadis Laboratory of Geodesy.
© 2008 McGraw-Hill Higher Education The Statistical Imagination Chapter 8. Parameter Estimation Using Confidence Intervals.
Neural Networks - Berrin Yanıkoğlu1 Applications and Examples From Mitchell Chp. 4.
Review of fundamental 1 Data mining in 1D: curve fitting by LLS Approximation-generalization tradeoff First homework assignment.
Overfitting, Bias/Variance tradeoff. 2 Content of the presentation Bias and variance definitions Parameters that influence bias and variance Bias and.
CORRELATION-REGULATION ANALYSIS Томский политехнический университет.
Lecture 2 Introduction to Neural Networks and Fuzzy Logic President UniversityErwin SitompulNNFL 2/1 Dr.-Ing. Erwin Sitompul President University
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Supervised Learning in ANNs
第 3 章 神经网络.
CSE 4705 Artificial Intelligence
The Calibration Process
Unfolding Problem: A Machine Learning Approach
Artificial Intelligence Chapter 3 Neural Networks
Nonlinear regression.
Crustal Deformation Analysis from Permanent GPS Networks
6.2 Grid Search of Chi-Square Space
5.2 Least-Squares Fit to a Straight Line
Artificial Intelligence Chapter 3 Neural Networks
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Bias-variance Trade-off
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
The loss function, the normal equation,
Artificial Intelligence Chapter 3 Neural Networks
Artificial Intelligence Chapter 3 Neural Networks
Computer Vision Lecture 19: Object Recognition III
Newton-Raphson Method
A Data Partitioning Scheme for Spatial Regression
Artificial Intelligence Chapter 3 Neural Networks
Presentation transcript:

A comparison of the ability of artificial neural network and polynomial fitting was carried out in order to model the horizontal deformation field. It is performed by means of the horizontal components of the GPS solutions in the Cascadia Subduction Zone. One set of the data is used to calculate the unknown parameters of the model and the other is used only for testing the accuracy of the model. The problem of overfitting (i.e., the substantial oscillation of the model between the training points) can be avoided by restricting the flexibility of the neural model. This can be done using an independent data set, namely the validation data, which has not been used to determine the parameters of the model. The proposed method is the so-called “stopped search method”, which can be used for obtaining a smooth and precise fitting model. However, when fitting high order polynomial, it is hard to overcome the negative effect of the overfitting problem. The computations are performed with Mathematica software, and the results are given in a symbolic form which can be used in the analysis of crustal deformation, e.g. strain analysis. Crustal velocity field modelling with neural network and polynomials Piroska Zaletnyik 1, Khosro Moghtased-Azar 2 1 Department of Geodesy and Surveying, Budapest University of Technology and Economics – Hungary, 2 Institut of Geodesy, University of Stuttgart – Germany, The adaptation of neural networks to the modeling of the deformation field offers geodesists a suitable tool for describing structural deformation. Overfitting problem can occur in higher order polynomials, but neural network overcomes the problem thanks to stopped search method. The greatest advantage of this method is that the solution can be given as an analytical function, which could be use to compute derivation of the velocity vectors for strain analysis. The first author wishes to thank to the Hungarian Eötvös Fellowship for supporting her visit at the Department of Geodesy and Geoinformatics of the University of Stuttgart (Germany), where this work has been accomplished. Abstract Results Fig 1.GPS determined horizontal velocity field by Pacific Northwest Geodetic Array (PANGA), which is plotted relative to North American Plate. Introduction Polynomial fitting When there are more points than the number of parameters, there is a possibility for adjustment calculation, i.e. for polynomial fitting. In this case 62 points were used for the adjustment. The calculations were carried out with Mathematica software. Figure 4 : Northing velocity model by polynomials Figure 5 : Northing velocity model by neural networks Conclusion Acknowledgement Neural network with stopped search method Testing set residuals (mm/year)minmaxmeanStd. Northing velocities Easting velocities ,86.1 Teaching set residuals (mm/year)minmaxmeanStd. Northing velocities Easting velocities Table1 :The statistics of the differences between polynomial model and real velocities in the 62 teaching points. Testing set residuals (mm/year)minmaxmeanStd. Northing velocities Easting velocities Teaching set residuals (mm/year)minmaxmeanStd. Northing velocities Easting velocities The GPS measurements to determine crustal strain rates were initiated in the Cascadia region (US Pacific Northwest and south-western British Columbia, Canada) more than a decade ago, with the first campaign measurements in 1986 and the establishment of permanent stations in Nowadays, continuous GPS data from the Pacific Northwest Geodetic Array processed by the geodesy laboratory serves as the data analysis facility for the Pacific Northwest Geodetic Array (PANGA). This organization has deployed an extensive network of continuous GPS sites that measure crustal deformation along the CSZ. Fig. 1 illustrates the horizontal velocity field along the Cascadia margin assuming the North American plate to be stable. Table 2. The statistics of the differences between 3D polynomial model and real velocities in the 20 testing points. The overfitting problem means that the error of the teaching set is decreasing while the error of the testing set is growing, in other words the network excessively fits the teaching points which is illustrated by Fig. 3. Overfitting problem Fig. 2. Overfitting problem Comparing the results of Table 1 and Table 2 we recognize a significant difference between the deviations of the teaching and the testing set. The determined model by polynomial, works well only in the teaching points but between them it does not work as well. The testing set, which was not used during the determination of the model, is also needed in order to qualify the results. As a classical approximation model, 3D polynomial fitting technique is used to build continuous velocity field as a function of geodetic coordinates. Displacement vector which can be derived from GPS observations have east, north and up components in topocentric coordinates. For modeling the horizontal displacement field we use only the north and the east elements. Accuracy of modeling is determined by differences between true values and values estimated by 3D polynomial fitting. When we increasee the degree of the polynomial, accuracy is increaseing up to the 6 th degree, but above that started to decrease, because of the deterioration of the conditions of the equations (ill conditioned equations). The 6 th order polynomial was the best fitting model. In this case 28 points are needed, because a two-variable 6 th order polynomial has 28 parameters. A central issue in choosing the most suitable model for a given problem is selecting the right structural complexity. Clearly, a model that contains too few parameters will not be flexible enough to approximate important features in the data. If the model contains too many parameters, it will approximate not only the data but also the noise in the data. Overfitting may be avoided by restricting the flexibility of the neural model in some way. The Neural Networks package in Mathematica offers a few ways to handle the overfitting problem. All solutions rely on the use of a second, independent data set, the so-called validation data, which has not been used to train the model. One way to handle this problem is the stopped search method. Stopped search refers to obtaining the network’s parameters at some intermediate iteration during the training process and not at the final iteration as it is normally done. During the training the values of the parameters are changing to reach the minimum of the mean square error (MSE). Using validation data, it is possible to identify an intermediate iteration where the parameter values yield a minimum MSE. At the end of the training process the parameter values at this minimum are the ones used in the delivered network model. In order to avoid the overfitting problem by means of stopped search method, we will need more data. A learning set and a validation set. Hence, we have to divide the used teaching set (62 points) into two sets, the first will be the learning set with 42 points and the remaining 20 points will be the validation set. In fig. 3 we can see the errors of the learning and the teaching set during the learning procedure of the neural network model for the northing velocities. The errors of the learning set (continuous line) decrease during the whole procedure, but the errors of the validation set (dashed line) are decreasing only until the 262 nd iteration step, from that point are growing. The maximum number of iteration was 500, but the best parameter set is the one calculated at the 262 nd iteration step. In the model in the hidden layer 7 neurons (nodes) were used. The selection of number of neurons is basically depends on the number of known points. In fact, by having more known data we can increase the numbers of neurons. Let us check the statistics of the residuals for the whole teaching set (62 points) in Table 3. Fig.3. Errors of learning (continuous line) and validation set (dashed line) during stopped search method Table 3. The statistics of the differences between neural network model and real velocities in the 62 teaching points. Table 4. The statistics of the differences between neural network model and real velocities in the 20 testing points. Let’s see the differences between the neural network model and the real velocities in the 20 testing points (Table 4.) Using neural network model with stopped search technique we can obtain a smooth and good fitting model, while in the case of high order polynomial model there are substantial oscillations between the teaching points. See fig. 4 and 5.