5.1 Introduction to Curve Fitting why do we fit data to a function?

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
Regression Analysis Using Excel. Econometrics Econometrics is simply the statistical analysis of economic phenomena Here, we just summarize some of the.
The General Linear Model. The Simple Linear Model Linear Regression.
Appendix to Chapter 1 Mathematics Used in Microeconomics © 2004 Thomson Learning/South-Western.
1 Learning Entity Specific Models Stefan Niculescu Carnegie Mellon University November, 2003.
Today Today: Chapter 9 Assignment: 9.2, 9.4, 9.42 (Geo(p)=“geometric distribution”), 9-R9(a,b) Recommended Questions: 9.1, 9.8, 9.20, 9.23, 9.25.
Ch. 6 The Normal Distribution
CHAPTER 6 Statistical Analysis of Experimental Data
Continuous Random Variables and Probability Distributions
Chi Square Distribution (c2) and Least Squares Fitting
Calibration & Curve Fitting
Chapter 4 Continuous Random Variables and Probability Distributions
SIMULATION MODELING AND ANALYSIS WITH ARENA
Least-Squares Regression
Physics 114: Lecture 15 Probability Tests & Linear Fitting Dale E. Gary NJIT Physics Department.
LINEAR REGRESSION Introduction Section 0 Lecture 1 Slide 1 Lecture 5 Slide 1 INTRODUCTION TO Modern Physics PHYX 2710 Fall 2004 Intermediate 3870 Fall.
Likelihood probability of observing the data given a model with certain parameters Maximum Likelihood Estimation (MLE) –find the parameter combination.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
Random Sampling, Point Estimation and Maximum Likelihood.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 Part 4 Curve Fitting.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
Graphs of Functions. Text Example SolutionThe graph of f (x) = x is, by definition, the graph of y = x We begin by setting up a partial table.
Modular 11 Ch 7.1 to 7.2 Part I. Ch 7.1 Uniform and Normal Distribution Recall: Discrete random variable probability distribution For a continued random.
Continuous Distributions The Uniform distribution from a to b.
R Introduction and Training Patrick Gurian, Drexel University CAMRA 1st QMRA Summer Institute August 7, 2006.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Expectation. Let X denote a discrete random variable with probability function p(x) (probability density function f(x) if X is continuous) then the expected.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Continuous Random Variables Chapter 6.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
Week 41 How to find estimators? There are two main methods for finding estimators: 1) Method of moments. 2) The method of Maximum likelihood. Sometimes.
NON-LINEAR REGRESSION Introduction Section 0 Lecture 1 Slide 1 Lecture 6 Slide 1 INTRODUCTION TO Modern Physics PHYX 2710 Fall 2004 Intermediate 3870 Fall.
Chapter 6 parameter estimation
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
Continuous Random Variables and Probability Distributions
CHAPTER 2.3 PROBABILITY DISTRIBUTIONS. 2.3 GAUSSIAN OR NORMAL ERROR DISTRIBUTION  The Gaussian distribution is an approximation to the binomial distribution.
Review of statistical modeling and probability theory Alan Moses ML4bio.
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and  2 Now, we need procedures to calculate  and  2, themselves.
Free Powerpoint Templates ROHANA BINTI ABDUL HAMID INSTITUT E FOR ENGINEERING MATHEMATICS (IMK) UNIVERSITI MALAYSIA PERLIS ROHANA BINTI ABDUL HAMID INSTITUT.
Introduction A probability distribution is obtained when probability values are assigned to all possible numerical values of a random variable. It may.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
CHAPTER 4 ESTIMATES OF MEAN AND ERRORS. 4.1 METHOD OF LEAST SQUARES I n Chapter 2 we defined the mean  of the parent distribution and noted that the.
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
Chapter 6 Continuous Random Variables Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Chapter 6 – Continuous Probability Distribution Introduction A probability distribution is obtained when probability values are assigned to all possible.
Data Modeling Patrice Koehl Department of Biological Sciences
Chapter 4: Basic Estimation Techniques
Kinetics of chemical reactions: overview
Physics 114: Lecture 13 Probability Tests & Linear Fitting
A First Course on Kinetics and Reaction Engineering
Basic Estimation Techniques
The Maximum Likelihood Method
The Normal Probability Distribution
Basic Estimation Techniques
Goodness-of-Fit Tests
3.1 Sums of Random Variables probability of z = x + y
2.3 Estimating PDFs and PDF Parameters
Continuous distributions
5.2 Least-Squares Fit to a Straight Line
5.4 General Linear Least-Squares
3.0 Functions of One Random Variable
6.1 Introduction to Chi-Square Space
Computing and Statistical Data Analysis / Stat 7
SKTN 2393 Numerical Methods for Nuclear Engineers
HKN ECE 313 Exam 2 Review Session
Continuous Distributions
The Normal Distribution
Presentation transcript:

5.1 Introduction to Curve Fitting why do we fit data to a function? the shape of probability distributions when curve fitting - normal, Poisson, and normal with scale changes the method of maximum likelihood as a strategy for optimizing parameters 5.1 : 1/10

Why Curve Fitting? (1) An equation parameter is the end goal of the experiment. As an example, consider a first order chemical reaction, A  B, where the concentration of A is measured as a function of time. C(t) = a0exp(-a1t) Let the dependent variable, C(t), have noise following a normal pdf. In such an experiment, the independent variable, t, is for all practical purposes noise-free. The goal of the experiment is to determine the rate constant, a1. To determine a0 and a1, the coefficients are adjusted until a "best fit" is found. For the example shown, the true values were a0 = 2.48 and a1 = 0.312. The best fit values were a0 = 2.60 and a1 = 0.33. a0 and a1 are dependent random variables. 5.1 : 2/10

Why Curve Fitting? (2) Curve fitting is an optimal strategy for testing multiple hypotheses. Consider the determination of molar absorptivity. The simplest strategy is to prepare a solution with an accurate concentration and measure the absorption several times. A preferred strategy is to prepare a series of solutions having different molar concentrations. A plot is then made of A versus C. For a 1-cm pathlength cell, the slope equals e. Ai = A0 + eCi Now two hypotheses can be tested in addition to determining the molar absorptivity. (1) A0 belongs to a pdf having 0 as its mean. This tests for the presence of unsuspected absorbers or instrumental offsets. (2) The data are described well by a straight line. This tests for chemical or instrumental failures of Beer's law. 5.1 : 3/10

f(x) = 2 + 0.5x + 0.3x2 (1) The contour graph at the left shows y as a function of x when g(y) is a normal pdf. The graph at the right shows g(y) for x' = 2 and x' = 8. Note that the width of the pdf is constant. The mean is given by f(x'). 5.1 : 4/10

f(x) = 2 + 0.5x + 0.3x2 (2) The contour graph at left shows y as a function of x when g(y) has a Poisson pdf. A common example would be data from a counting experiment. The graph at the right shows g(y) for x' = 2 and x' = 8. Note that the width of the pdf increases with x'. The mean is given by f(x'). 5.1 : 5/10

f(x) = 2 + 0.5x + 0.3x2 (3) The contour graph at left shows y as a function of x when g(y) is represented by two different normal pdfs. Over the range, 0  x  5, s = 2, while over the range, 5 < x  10, s = 1. The graph at the right shows g(y) for x' = 2 and x' = 8. The width of the pdf changes at x' = 5. This is supposed to reflect how noise might be reduced due to a decrease in instrumental sensitivity. 5.1 : 6/10

f(x) = 2 + 0.5x + 0.3x2 (4) In practice one measurement of y is made for each independently chosen value of x. The graph shows this for 11 values. The goal is to estimate the parameters of the equation. If there were no noise, the parameters could be determined by making three measurements and solving three simultaneous equations. When noise is present a probability-based method is used to estimate the parameters. Because of randomness in y, if a second set of 11 measurements are made, the estimates of the coefficients will vary randomly. 5.1 : 7/10

Probability for a Set of Values Consider the probability of obtaining a single measurement from a Poisson pdf. Now write the product representing the probability of observing N values from the same Poisson pdf. Expand the product in order to collect terms. The numerator can be converted into a summation and brought in front of the product operator. The result is the total probability written in a form more amenable to differentiation. 5.1 : 8/10

Method of Maximum Likelihood The Method of Maximum Likelihood asks the question - what definition of a pdf parameter will maximize the probability of observing the measured set of values? In this example we will ask what definition of m maximizes the probability. Start by taking the partial derivative with respect to the parameter (in this case m) and setting it equal to zero. We then see that the arithmetic average is the estimate of m that yields the highest probability for obtaining the measured set of data. 5.1 : 9/10

Example with m = 12 Two sets of 10 values were obtained from a Poisson pdf having a mean of 12. {10, 12, 18, 14, 16, 12, 16, 12, 14, 14} avg = 13.8 {14, 12, 9, 5, 12, 9, 6, 22,8, 8} avg = 10.5 The total probability graphs are shown below, where ptotal is computed as a function of trial m values. Note that the maximum probability does not correspond to the true value of 12! 5.1 : 10/10