3/2003 Rev 1 I.2.15-16 – slide 1 of 33 Session I.2.15-16 Part I Review of Fundamentals Module 2Basic Physics and Mathematics Used in Radiation Protection.

Slides:



Advertisements
Similar presentations
3/2003 Rev 1 I – slide 1 of 33 Session I Part I Review of Fundamentals Module 2Basic Physics and Mathematics Used in Radiation Protection.
Advertisements

3/2003 Rev 1 I.2.12 – slide 1 of 18 Part I Review of Fundamentals Module 2Basic Physics and Mathematics Used in Radiation Protection Session 12Statistics.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Regression and correlation methods
Lesson 10: Linear Regression and Correlation
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 ~ Curve Fitting ~ Least Squares Regression Chapter.
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Probabilistic & Statistical Techniques Eng. Tamer Eshtawi First Semester Eng. Tamer Eshtawi First Semester
Chapter 10 Curve Fitting and Regression Analysis
Ch11 Curve Fitting Dr. Deshi Ye
Chapter 10 Regression. Defining Regression Simple linear regression features one independent variable and one dependent variable, as in correlation the.
AP Statistics Chapters 3 & 4 Measuring Relationships Between 2 Variables.
The Islamic University of Gaza Faculty of Engineering Civil Engineering Department Numerical Analysis ECIV 3306 Chapter 17 Least Square Regression.
The Simple Regression Model
SIMPLE LINEAR REGRESSION
Correlation and Regression. Correlation What type of relationship exists between the two variables and is the correlation significant? x y Cigarettes.
SIMPLE LINEAR REGRESSION
Correlation and Regression Analysis
Linear Regression/Correlation
Relationships Among Variables
1 Chapter 10 Correlation and Regression We deal with two variables, x and y. Main goal: Investigate how x and y are related, or correlated; how much they.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS & Updated by SPIROS VELIANITIS.
Correlation & Regression
Linear Regression.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Simple Linear Regression Analysis Chapter 13.
SIMPLE LINEAR REGRESSION
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 ~ Curve Fitting ~ Least Squares Regression Chapter.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved Section 10-3 Regression.
Relationship of two variables
Chapter 15 Correlation and Regression
Chapter 8: Regression Analysis PowerPoint Slides Prepared By: Alan Olinsky Bryant University Management Science: The Art of Modeling with Spreadsheets,
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 Part 4 Curve Fitting.
Correlation and Regression Used when we are interested in the relationship between two variables. NOT the differences between means or medians of different.
1 1 Slide © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
Production Planning and Control. A correlation is a relationship between two variables. The data can be represented by the ordered pairs (x, y) where.
Examining Relationships in Quantitative Research
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
Basic Concepts of Correlation. Definition A correlation exists between two variables when the values of one are somehow associated with the values of.
4/2003 Rev 2 I.4.12 – slide 1 of 10 Session I.4.12 Part I Review of Fundamentals Module 4Sources of Radiation Session 12Filtration and Beam Quality IAEA.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
MARKETING RESEARCH CHAPTER 18 :Correlation and Regression.
3/2003 Rev 1 I.2.12 – slide 1 of 18 Part I Review of Fundamentals Module 2Basic Physics and Mathematics Used in Radiation Protection Session 12Statistics.
Correlation & Regression Analysis
4/2003 Rev 2 I.4.4 – slide 1 of 13 Session I.4.4 Part I Review of Fundamentals Module 4Sources of Radiation Session 4Sealed and Unsealed Sources and Isotope.
Regression Analysis. 1. To comprehend the nature of correlation analysis. 2. To understand bivariate regression analysis. 3. To become aware of the coefficient.
The “Big Picture” (from Heath 1995). Simple Linear Regression.
Lecture Slides Elementary Statistics Twelfth Edition
Linear Regression Essentials Line Basics y = mx + b vs. Definitions
Inference about the slope parameter and correlation
The simple linear regression model and parameter estimation
Chapter 4 Basic Estimation Techniques
Regression Analysis AGEC 784.
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Lecture Slides Elementary Statistics Thirteenth Edition
Correlation and Regression
CHAPTER 29: Multiple Regression*
Chapter 10 Correlation and Regression
Correlation and Simple Linear Regression
BA 275 Quantitative Business Methods
SIMPLE LINEAR REGRESSION
Simple Linear Regression and Correlation
Product moment correlation
SIMPLE LINEAR REGRESSION
Algebra Review The equation of a straight line y = mx + b
Created by Erin Hodgess, Houston, Texas
Multiple Regression Berlin Chen
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

3/2003 Rev 1 I – slide 1 of 33 Session I Part I Review of Fundamentals Module 2Basic Physics and Mathematics Used in Radiation Protection Session 15-16Data Analysis IAEA Post Graduate Educational Course Radiation Protection and Safety of Radiation Sources

3/2003 Rev 1 I – slide 2 of 33  Upon completion of this section the student should be able to demonstrate an understanding of the following statistical concepts pertaining to sample data:  Regression  Correlation Objectives

3/2003 Rev 1 I – slide 3 of 33  A “regression” is a statistical technique used to investigate the relationship among two or more variables  An independent variable, “x”, is linked to a dependent variable, “y”  The relationship may follow the generic form of a straight line, for example, y = mx + b, where m is the slope and b is the intercept Regression

3/2003 Rev 1 I – slide 4 of 33  Other forms of relationships include:  Y = a + bx + cx 2 a parabola  Y = ab x an exponential curve  Y = ax b a geometric curve Regression

3/2003 Rev 1 I – slide 5 of 33 Linear Least Squares Regression  Linear least squares regression is by far the most widely used modeling method. It is what most people mean when they say they have used "regression", "linear regression" or "least squares" to fit a model to their data.  Not only is linear least squares regression the most widely used modeling method, but it has been adapted to a broad range of situations that are outside its direct scope.  It plays a strong underlying role in many other modeling methods.

3/2003 Rev 1 I – slide 6 of 33 Linear Least Squares Regression Used directly, with an appropriate data set, linear least squares regression can be used to fit the data with any function of the form: F(x,  ) =  1 +  2 x 1 +  3 x 2 + … in which (see next slide)

3/2003 Rev 1 I – slide 7 of 33 Linear Least Squares Regression  Each explanatory variable in the function is multiplied by an unknown parameter  There is at most one unknown parameter with no corresponding explanatory variable  All of the individual terms are summed to produce the final function value

3/2003 Rev 1 I – slide 8 of 33 Linear Least Squares Regression  In statistical terms, any function that meets these criteria would be called a "linear function"  The term "linear" is used, even though the function may not be a straight line

3/2003 Rev 1 I – slide 9 of 33 Linear Least Squares Regression  The unknown parameters are considered to be variables and the explanatory variables are considered to be known coefficients corresponding to those "variables”  The problem becomes a system of “linear” equations that can be solved for the values of the unknown parameters

3/2003 Rev 1 I – slide 10 of 33 Linear Least Squares Regression  In the least squares method the unknown parameters are estimated by minimizing the sum of the squared deviations between the data and the model  The minimization process reduces the system of equations formed by the data to a sensible system of p (where p is the number of parameters in the functional part of the model) equations in p unknowns  This new system of equations is then solved to obtain the parameter estimates

3/2003 Rev 1 I – slide 11 of 33 Linear Least Squares Regression  Linear models are not limited to being straight lines or planes, but include a fairly wide range of shapes  For example, a simple quadratic curve is linear in the statistical sense  A straight-line model or a polynomial is also linear in the statistical sense because they are linear in the parameters, though not with respect to the observed explanatory variable, x

3/2003 Rev 1 I – slide 12 of 33 Linear Least Squares Regression  Just as models that are linear in the statistical sense do not have to be linear with respect to the explanatory variables, nonlinear models can be linear with respect to the explanatory variables, but not with respect to the parameters

3/2003 Rev 1 I – slide 13 of 33 Linear Least Squares Regression  For example, F(x,  ) =  1 +  1  2 x is linear in x but it cannot be written in the general form of a linear model  This is because the slope of this line is expressed as the product of two parameters  As a result, nonlinear least squares regression could be used to fit this model, but linear least squares cannot be used

3/2003 Rev 1 I – slide 14 of 33 Linear Least Squares Regression  Advantages:  Although there are types of data that are better described by functions that are nonlinear in the parameters, many processes in science and engineering are well-described by linear models  This is because either the processes are inherently linear or because, over short ranges, any process can be well-approximated by a linear model

3/2003 Rev 1 I – slide 15 of 33 Linear Least Squares Regression  Disadvantages:  The main disadvantages of linear least squares are:  limitations in the shapes that linear models can assume over long ranges  poor extrapolation properties  sensitivity to outliers  Linear models with nonlinear terms in the predictor variables curve relatively slowly, so for inherently nonlinear processes it becomes increasingly difficult to find a linear model that fits the data well as the range of the data increases.

3/2003 Rev 1 I – slide 16 of 33 Linear Least Squares Regression  Finally, while the method of least squares often gives optimal estimates of the unknown parameters, it is very sensitive to the presence of unusual data points in the data used to fit a model  One or two outliers can sometimes seriously skew the results of a least squares analysis

3/2003 Rev 1 I – slide 17 of 33 Linear Least Squares Regression  Application of this concept enables us to model a relationship to data  Do not expect the data to perfectly follow the model  The idea is to establish a model that will provide a “best fit” in representing the data

3/2003 Rev 1 I – slide 18 of 33 Regression Example For the following data, what is the function that relates the “Y” values to the “X” values? X12357 Y43489

3/2003 Rev 1 I – slide 19 of 33 Plot of sample data Regression Example

3/2003 Rev 1 I – slide 20 of 33  A line is drawn through the data points  The difference between the value represented by the line and the observed value is determined  This value is squared (which gets rid of negative values) Regression Example

3/2003 Rev 1 I – slide 21 of 33 Sum of Squares - Total  The total sum of the squares is calculated for each data point. This is called the “sum of squares,” or SS  The line that is the “best fit” is one that has the smallest SS value SS TOTAL =  (Y i – Y AVG ) 2

3/2003 Rev 1 I – slide 22 of 33 Sum of Squares - Residual The residual sum of the squares, SS RES, is calculated by determining all the residuals, squaring them, and summing the squares. It has n-2 degrees of freedom SS RES =  (Y i – Ý i ) 2 where Ý i is the “fitted” or predicted value of Y Another procedure for calculating SS RES is: SS RES = SS TOTAL - SS REG

3/2003 Rev 1 I – slide 23 of 33 Sum of Squares - Regression The regression sum of the squares, SS REG, is calculated by determining all the residuals, squaring them, and summing the squares. It has 1 degree of freedom. SS REG =  (Ý i - Y AVG ) 2 where Ý i is the “fitted” or predicted value of Y

3/2003 Rev 1 I – slide 24 of 33 F Statistic  When comparing the variance of two different groups (populations), the null hypothesis, H 0, is that the variance of the two groups is equal (  A 2 =  B 2 )  The alternative hypothesis, H 1, is that the variance of the two groups is not equal (  A 2   B 2 )

3/2003 Rev 1 I – slide 25 of 33 F Statistic  From the first population, you make n A observations and calculate the sample variance, S A 2 with df A = n A –1  From the second population, you make n B observations and calculate the sample variance, S B 2 with df B = n B –1

3/2003 Rev 1 I – slide 26 of 33 F Statistic  Let S MAX 2 and S MIN 2 denote the larger and smaller of S A 2 and S B 2, respectively  Likewise, let df MAX and df MIN denote their respective degrees of freedom  The test statistic, F, also known as the “F ratio” or “variance ratio” is: F = S max 2 S min 2

3/2003 Rev 1 I – slide 27 of 33 F Statistic  If H 0 is correct, the F ratio should not be much larger than 1  The question is “how large is large?”  The quartile (for a 2 tail distribution) is f (df MAX, df MIN )

3/2003 Rev 1 I – slide 28 of 33 Source of Variation Sum of Squares Degrees of Freedom Mean Square F Regression SS REG 1 MS REG = SS REG /df MS REG MS RES Residual SS RES n - 2 MS RES = SS RES /df Total SS TOTAL n - 1 n - 1 Regression Analysis Table

3/2003 Rev 1 I – slide 29 of 33 Slope  The equation for the slope of a regression line is: B =  For the data in the example, the slope, B = 1.04 [n  X 2 – (  X) 2 ] [n  (XY) – (  X)(  Y)]

3/2003 Rev 1 I – slide 30 of 33 Regression analysis plot of sample data Regression Analysis

3/2003 Rev 1 I – slide 31 of 33 A regression analysis of these sample data indicates the slope to be 1.04, with an intercept value of 1.84 (see next slide) Regression Analysis

3/2003 Rev 1 I – slide 32 of 33 Regression Analysis

3/2003 Rev 1 I – slide 33 of 33 Where to Get More Information  Cember, H., Johnson, T. E., Introduction to Health Physics, 4th Edition, McGraw-Hill, New York (2008)  Martin, A., Harbison, S. A., Beach, K., Cole, P., An Introduction to Radiation Protection, 6 th Edition, Hodder Arnold, London (2012)  Firestone, R.B., Baglin, C.M., Frank-Chu, S.Y., Eds., Table of Isotopes (8 th Edition, 1999 update), Wiley, New York (1999)