CHAPTER 3 Model Fitting. Introduction Possible tasks when analyzing a collection of data points: Fitting a selected model type or types to the data Choosing.

Slides:



Advertisements
Similar presentations
Simple Linear Regression Analysis
Advertisements

11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 ~ Curve Fitting ~ Least Squares Regression Chapter.
Chapter 7 Statistical Data Treatment and Evaluation
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
11 Simple Linear Regression and Correlation CHAPTER OUTLINE
Chapter 10 Curve Fitting and Regression Analysis
P M V Subbarao Professor Mechanical Engineering Department
Regression Analysis Using Excel. Econometrics Econometrics is simply the statistical analysis of economic phenomena Here, we just summarize some of the.
Read Chapter 17 of the textbook
ECIV 201 Computational Methods for Civil Engineers Richard P. Ray, Ph.D., P.E. Error Analysis.
Curve Fitting and Interpolation: Lecture (IV)
Least Square Regression
Curve-Fitting Regression
Least Square Regression
The Islamic University of Gaza Faculty of Engineering Civil Engineering Department Numerical Analysis ECIV 3306 Chapter 17 Least Square Regression.
1 Simple Linear Regression Chapter Introduction In this chapter we examine the relationship among interval variables via a mathematical equation.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 171 CURVE.
Introduction to Linear and Logistic Regression. Basic Ideas Linear Transformation Finding the Regression Line Minimize sum of the quadratic residuals.
Lecture 17 Interaction Plots Simple Linear Regression (Chapter ) Homework 4 due Friday. JMP instructions for question are actually for.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 171 Least.
Introduction to Regression Analysis, Chapter 13,
Simple Linear Regression. Introduction In Chapters 17 to 19, we examine the relationship between interval variables via a mathematical equation. The motivation.
Calibration & Curve Fitting
Lecture 3: Bivariate Data & Linear Regression 1.Introduction 2.Bivariate Data 3.Linear Analysis of Data a)Freehand Linear Fit b)Least Squares Fit c)Interpolation/Extrapolation.
Lecture 3-2 Summarizing Relationships among variables ©
1 Doing Statistics for Business Doing Statistics for Business Data, Inference, and Decision Making Marilyn K. Pelosi Theresa M. Sandifer Chapter 11 Regression.
Least-Squares Regression
CpE- 310B Engineering Computation and Simulation Dr. Manal Al-Bzoor
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 ~ Curve Fitting ~ Least Squares Regression Chapter.
Introduction to Linear Regression and Correlation Analysis
Correlation and Regression
Ch 8.1 Numerical Methods: The Euler or Tangent Line Method
1 Statistical Analysis - Graphical Techniques Dr. Jerrell T. Stracener, SAE Fellow Leadership in Engineering EMIS 7370/5370 STAT 5340 : PROBABILITY AND.
Chapter 13 Statistics © 2008 Pearson Addison-Wesley. All rights reserved.
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 Part 4 Curve Fitting.
Ch4 Describing Relationships Between Variables. Pressure.
Linear Regression James H. Steiger. Regression – The General Setup You have a set of data on two variables, X and Y, represented in a scatter plot. You.
© 2008 Pearson Addison-Wesley. All rights reserved Chapter 1 Section 13-6 Regression and Correlation.
MECN 3500 Inter - Bayamon Lecture 9 Numerical Methods for Engineering MECN 3500 Professor: Dr. Omar E. Meza Castillo
Chapter 8 Curve Fitting.
Curve-Fitting Regression
MECN 3500 Inter - Bayamon Lecture 3 Numerical Methods for Engineering MECN 3500 Professor: Dr. Omar E. Meza Castillo
Chapter 11 Linear Regression Straight Lines, Least-Squares and More Chapter 11A Can you pick out the straight lines and find the least-square?
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
MA3264 Mathematical Modelling Lecture 3 Model Fitting.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
1 Data Analysis Linear Regression Data Analysis Linear Regression Ernesto A. Diaz Department of Mathematics Redwood High School.
MODEL FITTING jiangyushan. Introduction The goal of model fitting is to choose values for the parameters in a function to best describe a set of data.
Curve Fitting Pertemuan 10 Matakuliah: S0262-Analisis Numerik Tahun: 2010.
Chapter 3- Model Fitting. Three Tasks When Analyzing Data: 1.Fit a model type to the data. 2.Choose the most appropriate model from the ones that have.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Simple Linear Regression Analysis Chapter 13.
Curve Fitting Introduction Least-Squares Regression Linear Regression Polynomial Regression Multiple Linear Regression Today’s class Numerical Methods.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 ~ Curve Fitting ~ Least Squares Regression.
1 Statistical Analysis - Graphical Techniques Dr. Jerrell T. Stracener, SAE Fellow Leadership in Engineering EMIS 7370/5370 STAT 5340 : PROBABILITY AND.
The simple linear regression model and parameter estimation
Department of Mathematics
Chapter 4 Basic Estimation Techniques
Fitting Equations to Data
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Part 5 - Chapter
Part 5 - Chapter 17.
10.2 Regression If the value of the correlation coefficient is significant, the next step is to determine the equation of the regression line which is.
Part 5 - Chapter 17.
Descriptive Analysis and Presentation of Bivariate Data
Undergraduated Econometrics
M248: Analyzing data Block D UNIT D2 Regression.
Least Square Regression
Presentation transcript:

CHAPTER 3 Model Fitting

Introduction Possible tasks when analyzing a collection of data points: Fitting a selected model type or types to the data Choosing the most appropriate model from competing types that have been fitted. For example, we may need to determine whether the best- fitting exponential model is a better model than the best-fitting polynomial model. Making predictions from the collected data.

Relationship Between Model Fitting and Interpolation The modeler tends to emphasize the proposed models over the data when model fitting, whereas when interpolating, he or she places greater confidence in the collected data and attaches less significance to the form of the model. In a sense, explicative models are theory driven, whereas predictive models are data driven.

Sources of Error in the Modeling Process Formulation Errors Assumption that certain variables are negligible. simplifications in describing interrelationships among the variables in the various submodels. Truncation Errors Attributable to the numerical method used to solve a mathematical problem. For example, we may find it necessary to approximate sin x with a polynomial representation obtained from the power series An error will be introduced when the series is truncated to produce the polynomial.

Sources of Error in the Modeling Process Round-off Errors Caused by using a finite digit machine for computation. When many arithmetic operations are performed in succession, each with its own round-off, the accumulated effect of round-off can significantly alter the numbers that are supposed to be the answer. Measurement Errors Caused by imprecision in the data collection. This imprecision may include such diverse things as human errors in recording or reporting the data or the actual physical limitations of the laboratory equipment.

Fitting Models to Data Graphically The modeler needs at least as many data points as there are arbitrary constants in the model curve. The spacing of the data points within the interval of study is important because any part of the interval over which the model must fit particularly well can be weighted by using unequal spacing. It is useful to think of each data point as an interval of relative confidence rather than as a single point.

Visual Model Fitting with the Original Data Suppose we want to fit the model y = a x + b to the data shown in the figure of the previous slide Minimizing the sum of the absolute deviations from the fitted line Minimizing the largest absolute deviation from the fitted line

Example What criteria would you use to fit this data set

Transforming the Data Example Suppose, for example, that a relationship of the form y = Ce x is suspected for some submodel and the data shown in table have been collected. Plot y versus e x From the figure, the slope of the line is approximated as

Transforming the Data An alternative technique: Take the logarithm of each side of the equation y = Ce x to obtain ln y = ln C + x The intercept ln C is approximately 1.1, giving C = e 1.1 ≈ 3.0 as before.

3.2 Analytic Methods of Model Fitting Chebyshev Approximation Criterion Given a collection of m data points (x i, y i ), i = 1, 2, …, m, fit the collection to the line y = m x + b, determined by the parameters a and b, that minimizes the distance between any data point (x i, y i ) and its corresponding data point on the line (x i, ax i + b). That is, minimize the largest absolute deviation |y i − y(x i )| over the entire collection of data points. Linear Programming Simplex Method Minimizing the Sum of the Absolute Deviations

Analytic Methods of Model Fitting Least Squares Criterion. Proof of the Minimization of the Mean Squared Residual (MSR)

3.3 Applying the Least-Squares Criterion Fitting a Straight Line Suppose a model of the form y = Ax + B Denote the least-squares estimate of y = Ax + B by y = ax + b. Applying the least-squares criterion to this situation requires the minimization of A necessary condition for optimality is that the two partial derivatives ∂S/∂a and ∂S/∂b equal zero, yielding the equations

Applying the Least-Squares Criterion for a Straight Line Show that The slope is given by The y intercept is given by These equations are called the Normal Equations.

Fitting a Power Curve Use the least-squares criterion to fit a curve of the form y=Ax n, where n is fixed, to a given collection of data points. Application of the criterion requires minimization of A necessary condition for optimality is that the derivative ds/da equal zero, giving the equation

Fitting a Power Curve Solving for a: Example Fit y=Ax2 to the following data xy

Transformed Least-Squares Fit Suppose we wish to fit the power curve y=Ax N to a collection of data points. Let’s denote the estimate of A by  and the estimate of N by n. Taking the logarithm of both sides of the equation y =  x n yields Since the resulting graph is a straight line, we can use the Normal Equations (for the same data set in the example)

Transformed Least-Squares Fit

Show that the least-squares model results in y = x 2 The least-squares best fit to the transformed equations does not coincide with the least-squares best fit of the original equations. The reason for this discrepancy is that the resulting optimization problems are different. In the case of the original problem, we are finding the curve that minimizes the sum of the squares of the deviations using the original data, whereas in the case of the transformed problem, we are minimizing the sum of the squares of the deviations using the transformed variables.

Choosing the Best Model Which of the last three models works best? It is tempting to choose the model that has the smallest sum of squared deviations for the given data set. However, in isolation these indicators may be very misleading. For example, consider the data displayed in the figure. In all of the four cases, the model y = x results in exactly the same sum of squared deviations. Without the benefit of the graphs, therefore, we might conclude that in each case the model fits the data about the same. However, as the graphs show, there is a significant variation in each model’s ability to capture the trend of the data.

Homework (Due Wed 09/26/12) Page 86 Problem # 4 Page 94 Problem # 2 Page 108 Problems #4, 6 (estimate the parameters not just graphically but using all the tools learned in chapter 3) Page 120 Problem # 2 (skip the computation of D, d max, and bound for c max ) Problems #3, 4