Econ 482 Lecture 1 I. Administration: Introduction Syllabus Thursday, Jan 16 th, “Lab” class is from 5-6pm in Savery 117 II. Material: Start of Statistical.

Slides:



Advertisements
Similar presentations
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: expected value of a random variable Original citation: Dougherty,
Advertisements

CS433: Modeling and Simulation
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Confidence Intervals Chapter 8.
Probability Distributions CSLU 2850.Lo1 Spring 2008 Cameron McInally Fordham University May contain work from the Creative Commons.
Business Statistics for Managerial Decision
Discrete Random Variables
1 PROBABILITY DISTRIBUTION EXAMPLE: X IS THE SUM OF TWO DICE red This sequence provides an example of a discrete random variable. Suppose that you.
Lecture 2 Today: Statistical Review cont’d:
ASYMPTOTIC PROPERTIES OF ESTIMATORS: PLIMS AND CONSISTENCY
Point estimation, interval estimation
BCOR 1020 Business Statistics Lecture 9 – February 14, 2008.
Probability and Probability Distributions
Random Variables and Distributions Lecture 5: Stat 700.
Continuous Random Variables and Probability Distributions
QMS 6351 Statistics and Research Methods Probability and Probability distributions Chapter 4, page 161 Chapter 5 (5.1) Chapter 6 (6.2) Prof. Vera Adamchik.
Variance Fall 2003, Math 115B. Basic Idea Tables of values and graphs of the p.m.f.’s of the finite random variables, X and Y, are given in the sheet.
Copyright © Cengage Learning. All rights reserved. 4 Continuous Random Variables and Probability Distributions.
EXPECTED VALUE OF A RANDOM VARIABLE 1 The expected value of a random variable, also known as its population mean, is the weighted average of its possible.
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: expected value of a function of a random variable Original citation:
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: continuous random variables Original citation: Dougherty, C. (2012)
EXERCISE R.4 1 R.4*Find the expected value of X in Exercise R.2. [R.2*A random variable X is defined to be the larger of the numbers when two dice are.
1 CONTINUOUS RANDOM VARIABLES A discrete random variable is one that can take only a finite set of values. The sum of the numbers when two dice are thrown.
Chapter 7: Random Variables
Chapter 7: The Normal Probability Distribution
Chapter 5 Discrete Random Variables and Probability Distributions ©
Chapter 6: Probability Distributions
McGraw-Hill/Irwin Copyright © 2013 by The McGraw-Hill Companies, Inc. All rights reserved.
Probability The definition – probability of an Event Applies only to the special case when 1.The sample space has a finite no.of outcomes, and 2.Each.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 11 Section 1 – Slide 1 of 34 Chapter 11 Section 1 Random Variables.
QBM117 Business Statistics Probability and Probability Distributions Continuous Probability Distributions 1.
K. Shum Lecture 16 Description of random variables: pdf, cdf. Expectation. Variance.
Applied Business Forecasting and Regression Analysis Review lecture 2 Randomness and Probability.
Random Variables Numerical Quantities whose values are determine by the outcome of a random experiment.
1 G Lect 3b G Lecture 3b Why are means and variances so useful? Recap of random variables and expectations with examples Further consideration.
LECTURER PROF.Dr. DEMIR BAYKA AUTOMOTIVE ENGINEERING LABORATORY I.
Chapter 5 Discrete Probability Distributions
Mean and Standard Deviation of Discrete Random Variables.
The two way frequency table The  2 statistic Techniques for examining dependence amongst two categorical variables.
Discrete Random Variables
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Mean, Variance, Moments and.
Chapter 8 Continuous Probability Distributions Sir Naseer Shahzada.
CY1B2 Statistics1 (ii) Poisson distribution The Poisson distribution resembles the binomial distribution if the probability of an accident is very small.
Estimators and estimates: An estimator is a mathematical formula. An estimate is a number obtained by applying this formula to a set of sample data. 1.
Chapter 5: Measures of Dispersion. Dispersion or variation in statistics is the degree to which the responses or values obtained from the respondents.
President UniversityErwin SitompulPBST 3/1 Dr.-Ing. Erwin Sitompul President University Lecture 3 Probability and Statistics
Probability Theory Modelling random phenomena. Permutations the number of ways that you can order n objects is: n! = n(n-1)(n-2)(n-3)…(3)(2)(1) Definition:
1 7.5 CONTINUOUS RANDOM VARIABLES Continuous data occur when the variable of interest can take on anyone of an infinite number of values over some interval.
Random Variables. Numerical Outcomes Consider associating a numerical value with each sample point in a sample space. (1,1) (1,2) (1,3) (1,4) (1,5) (1,6)
Continuous Random Variables and Probability Distributions
Definition of, the expected value of a function of X : 1 EXPECTED VALUE OF A FUNCTION OF A RANDOM VARIABLE To find the expected value of a function of.
Chapter 20 Statistical Considerations Lecture Slides The McGraw-Hill Companies © 2012.
Review Know properties of Random Variables
Measurements and Their Analysis. Introduction Note that in this chapter, we are talking about multiple measurements of the same quantity Numerical analysis.
CHAPTER – 1 UNCERTAINTIES IN MEASUREMENTS. 1.3 PARENT AND SAMPLE DISTRIBUTIONS  If we make a measurement x i in of a quantity x, we expect our observation.
7.2 Means & Variances of Random Variables AP Statistics.
Copyright © Cengage Learning. All rights reserved. 4 Continuous Random Variables and Probability Distributions.
Econ 482.
Introduction to Econometrics, 5th edition
Discrete Probability Distributions
Mean & Variance of a Distribution
CONTINUOUS RANDOM VARIABLES AND THE NORMAL DISTRIBUTION
Virtual University of Pakistan
AP Statistics Chapter 16 Notes.
CHAPTER – 1.2 UNCERTAINTIES IN MEASUREMENTS.
Chapter 5: Discrete Probability Distributions
Random Variables and Probability Distributions
Discrete Random Variables and Probability Distributions
Normal Probability Distribution Lecture 1 Sections: 6.1 – 6.2
Presentation transcript:

Econ 482 Lecture 1 I. Administration: Introduction Syllabus Thursday, Jan 16 th, “Lab” class is from 5-6pm in Savery 117 II. Material: Start of Statistical Review: Discrete and continuous random variables Expected value, variance

Random variables: A random variable is any variable whose value cannot be predicted exactly 1.Discrete random variables: Random variable which has a specific (countable set) of possible values Examples? 2. Continuous random variables: Random variable which can take any value of a continuous range of values

Properties of discrete random variables

© Christopher Dougherty 1999–2006 PROBABILITY DISTRIBUTION EXAMPLE: X IS THE SUM OF TWO DICE red This sequence provides an example of a discrete random variable. Suppose that you have a red die which, when thrown, takes the numbers from 1 to 6 with equal probability.

© Christopher Dougherty 1999–2006 red green Suppose that you also have a green die that can take the numbers from 1 to 6 with equal probability. PROBABILITY DISTRIBUTION EXAMPLE: X IS THE SUM OF TWO DICE

© Christopher Dougherty 1999–2006 red green We will define a random variable X as the sum of the numbers when the dice are thrown. PROBABILITY DISTRIBUTION EXAMPLE: X IS THE SUM OF TWO DICE

© Christopher Dougherty 1999–2006 For example, if the red die is 4 and the green one is 6, X is equal to 10. red green PROBABILITY DISTRIBUTION EXAMPLE: X IS THE SUM OF TWO DICE

© Christopher Dougherty 1999–2006 red green Similarly, if the red die is 2 and the green one is 5, X is equal to 7. PROBABILITY DISTRIBUTION EXAMPLE: X IS THE SUM OF TWO DICE

© Christopher Dougherty 1999–2006 red green The table shows all the possible outcomes. PROBABILITY DISTRIBUTION EXAMPLE: X IS THE SUM OF TWO DICE

© Christopher Dougherty 1999–2006 red green X If you look at the table, you can see that X can be any of the numbers from 2 to 12. PROBABILITY DISTRIBUTION EXAMPLE: X IS THE SUM OF TWO DICE

© Christopher Dougherty 1999–2006 red green Xf We will now define f, the frequencies associated with the possible values of X. PROBABILITY DISTRIBUTION EXAMPLE: X IS THE SUM OF TWO DICE

© Christopher Dougherty 1999–2006 red green Xf For example, there are four outcomes which make X equal to 5. PROBABILITY DISTRIBUTION EXAMPLE: X IS THE SUM OF TWO DICE

© Christopher Dougherty 1999–2006 red green Xf Similarly you can work out the frequencies for all the other values of X. PROBABILITY DISTRIBUTION EXAMPLE: X IS THE SUM OF TWO DICE

© Christopher Dougherty 1999–2006 red green Xfp Finally we will derive the probability of obtaining each value of X. PROBABILITY DISTRIBUTION EXAMPLE: X IS THE SUM OF TWO DICE

© Christopher Dougherty 1999–2006 red green Xfp If there is 1/6 probability of obtaining each number on the red die, and the same on the green die, each outcome in the table will occur with 1/36 probability. PROBABILITY DISTRIBUTION EXAMPLE: X IS THE SUM OF TWO DICE

© Christopher Dougherty 1999–2006 red green Xfp 211/36 322/36 433/36 544/36 655/36 766/36 855/36 944/ / / /36 Hence to obtain the probabilities associated with the different values of X, we divide the frequencies by 36. PROBABILITY DISTRIBUTION EXAMPLE: X IS THE SUM OF TWO DICE

© Christopher Dougherty 1999–2006 The distribution is shown graphically. in this example it is symmetrical, highest for X equal to 7 and declining on either side. 6 __ 36 5 __ 36 4 __ 36 3 __ 36 2 __ 36 2 __ 36 3 __ 36 5 __ 36 4 __ 36 probability X PROBABILITY DISTRIBUTION EXAMPLE: X IS THE SUM OF TWO DICE

© Christopher Dougherty 1999–2006 Definition of E(X), the expected value of X: EXPECTED VALUE OF A RANDOM VARIABLE The expected value of a random variable, also known as its population mean, is the weighted average of its possible values, the weights being the probabilities attached to the values.

© Christopher Dougherty 1999–2006 Definition of E(X), the expected value of X: EXPECTED VALUE OF A RANDOM VARIABLE Note that the sum of the probabilities must be unity, so there is no need to divide by the sum of the weights.

© Christopher Dougherty 1999–2006 x i x 1 x 2 x 3 x 4 x 5 x 6 x 7 x 8 x 9 x 10 x 11 EXPECTED VALUE OF A RANDOM VARIABLE This sequence shows how the expected value is calculated, first in abstract and then with the random variable defined in the first sequence. We begin by listing the possible values of X.

© Christopher Dougherty 1999–2006 x i p i x 1 p 1 x 2 p 2 x 3 p 3 x 4 p 4 x 5 p 5 x 6 p 6 x 7 p 7 x 8 p 8 x 9 p 9 x 10 p 10 x 11 p 11 EXPECTED VALUE OF A RANDOM VARIABLE Next we list the probabilities attached to the different possible values of X.

© Christopher Dougherty 1999–2006 x i p i x 1 p 1 x 2 p 2 x 3 p 3 x 4 p 4 x 5 p 5 x 6 p 6 x 7 p 7 x 8 p 8 x 9 p 9 x 10 p 10 x 11 p 11 EXPECTED VALUE OF A RANDOM VARIABLE Then we define a column in which the values are weighted by the corresponding probabilities.

© Christopher Dougherty 1999–2006 x i p i x 1 p 1 x 2 p 2 x 3 p 3 x 4 p 4 x 5 p 5 x 6 p 6 x 7 p 7 x 8 p 8 x 9 p 9 x 10 p 10 x 11 p 11 EXPECTED VALUE OF A RANDOM VARIABLE We do this for each value separately.

© Christopher Dougherty 1999–2006 x i p i x 1 p 1 x 2 p 2 x 3 p 3 x 4 p 4 x 5 p 5 x 6 p 6 x 7 p 7 x 8 p 8 x 9 p 9 x 10 p 10 x 11 p 11 EXPECTED VALUE OF A RANDOM VARIABLE Here we are assuming that n, the number of possible values, is equal to 11, but it could be any number.

© Christopher Dougherty 1999–2006 x i p i x 1 p 1 x 2 p 2 x 3 p 3 x 4 p 4 x 5 p 5 x 6 p 6 x 7 p 7 x 8 p 8 x 9 p 9 x 10 p 10 x 11 p 11  x i p i = E(X) EXPECTED VALUE OF A RANDOM VARIABLE The expected value is the sum of the entries in the third column.

© Christopher Dougherty 1999–2006 x i p i x i p i x i p i x 1 p 1 x 1 p 1 21/36 x 2 p 2 x 2 p 2 32/36 x 3 p 3 x 3 p 3 43/36 x 4 p 4 x 4 p 4 54/36 x 5 p 5 x 5 p 5 65/36 x 6 p 6 x 6 p 6 76/36 x 7 p 7 x 7 p 7 85/36 x 8 p 8 x 8 p 8 94/36 x 9 p 9 x 9 p 9 103/36 x 10 p 10 x 10 p /36 x 11 p 11 x 11 p /36  x i p i = E(X) EXPECTED VALUE OF A RANDOM VARIABLE The random variable X defined in the previous sequence could be any of the integers from 2 to 12 with probabilities as shown.

© Christopher Dougherty 1999–2006 x i p i x i p i x 1 p 1 x 1 p 1 21/362/36 x 2 p 2 x 2 p 2 32/36 x 3 p 3 x 3 p 3 43/36 x 4 p 4 x 4 p 4 54/36 x 5 p 5 x 5 p 5 65/36 x 6 p 6 x 6 p 6 76/36 x 7 p 7 x 7 p 7 85/36 x 8 p 8 x 8 p 8 94/36 x 9 p 9 x 9 p 9 103/36 x 10 p 10 x 10 p /36 x 11 p 11 x 11 p /36  x i p i = E(X) EXPECTED VALUE OF A RANDOM VARIABLE X could be equal to 2 with probability 1/36, so the first entry in the calculation of the expected value is 2/36.

© Christopher Dougherty 1999–2006 x i p i x i p i x 1 p 1 x 1 p 1 21/362/36 x 2 p 2 x 2 p 2 32/366/36 x 3 p 3 x 3 p 3 43/36 x 4 p 4 x 4 p 4 54/36 x 5 p 5 x 5 p 5 65/36 x 6 p 6 x 6 p 6 76/36 x 7 p 7 x 7 p 7 85/36 x 8 p 8 x 8 p 8 94/36 x 9 p 9 x 9 p 9 103/36 x 10 p 10 x 10 p /36 x 11 p 11 x 11 p /36  x i p i = E(X) EXPECTED VALUE OF A RANDOM VARIABLE The probability of x being equal to 3 was 2/36, so the second entry is 6/36.

© Christopher Dougherty 1999–2006 x i p i x i p i x 1 p 1 x 1 p 1 21/362/36 x 2 p 2 x 2 p 2 32/366/36 x 3 p 3 x 3 p 3 43/3612/36 x 4 p 4 x 4 p 4 54/3620/36 x 5 p 5 x 5 p 5 65/3630/36 x 6 p 6 x 6 p 6 76/3642/36 x 7 p 7 x 7 p 7 85/3640/36 x 8 p 8 x 8 p 8 94/3636/36 x 9 p 9 x 9 p 9 103/3630/36 x 10 p 10 x 10 p /3622/36 x 11 p 11 x 11 p /3612/36  x i p i = E(X) EXPECTED VALUE OF A RANDOM VARIABLE Similarly for the other 9 possible values.

© Christopher Dougherty 1999–2006 x i p i x i p i x 1 p 1 x 1 p 1 21/362/36 x 2 p 2 x 2 p 2 32/366/36 x 3 p 3 x 3 p 3 43/3612/36 x 4 p 4 x 4 p 4 54/3620/36 x 5 p 5 x 5 p 5 65/3630/36 x 6 p 6 x 6 p 6 76/3642/36 x 7 p 7 x 7 p 7 85/3640/36 x 8 p 8 x 8 p 8 94/3636/36 x 9 p 9 x 9 p 9 103/3630/36 x 10 p 10 x 10 p /3622/36 x 11 p 11 x 11 p /3612/36  x i p i = E(X) 252/36 To obtain the expected value, we sum the entries in this column. EXPECTED VALUE OF A RANDOM VARIABLE

© Christopher Dougherty 1999–2006 The expected value turns out to be 7. Actually, this was obvious anyway. We saw in the previous sequence that the distribution is symmetrical about 7. EXPECTED VALUE OF A RANDOM VARIABLE x i p i x i p i x 1 p 1 x 1 p 1 21/362/36 x 2 p 2 x 2 p 2 32/366/36 x 3 p 3 x 3 p 3 43/3612/36 x 4 p 4 x 4 p 4 54/3620/36 x 5 p 5 x 5 p 5 65/3630/36 x 6 p 6 x 6 p 6 76/3642/36 x 7 p 7 x 7 p 7 85/3640/36 x 8 p 8 x 8 p 8 94/3636/36 x 9 p 9 x 9 p 9 103/3630/36 x 10 p 10 x 10 p /3622/36 x 11 p 11 x 11 p /3612/36  x i p i = E(X) 252/36 = 7

© Christopher Dougherty 1999–2006 Alternative notation for E(X): E(X) =  X Very often the expected value of a random variable is represented by , the Greek m. If there is more than one random variable, their expected values are differentiated by adding subscripts to . EXPECTED VALUE OF A RANDOM VARIABLE

© Christopher Dougherty 1999–2006 Definition of E[g(X)], the expected value of a function of X: To find the expected value of a function of a random variable, you calculate all the possible values of the function, weight them by the corresponding probabilities, and sum the results. EXPECTED VALUE OF A FUNCTION OF A RANDOM VARIABLE

© Christopher Dougherty 1999–2006 Definition of E[g(X)], the expected value of a function of X: Example: For example, the expected value of X 2 is found by calculating all its possible values, multiplying them by the corresponding probabilities, and summing. EXPECTED VALUE OF A FUNCTION OF A RANDOM VARIABLE

© Christopher Dougherty 1999–2006 xipix1p1x2p2x3p3……………………………………xnpn xipix1p1x2p2x3p3……………………………………xnpn ………………… First you list the possible values of X and the corresponding probabilities. EXPECTED VALUE OF A FUNCTION OF A RANDOM VARIABLE

© Christopher Dougherty 1999–2006 x i p i g(x i ) x 1 p 1 g(x 1 ) x 2 p 2 g(x 2 ) x 3 p 3 g(x 3 ) ………... x n p n g(x n ) Next you calculate the function of X for each possible value of X. EXPECTED VALUE OF A FUNCTION OF A RANDOM VARIABLE

© Christopher Dougherty 1999–2006 x i p i g(x i ) g(x i ) p i x 1 p 1 g(x 1 )g(x 1 ) p 1 x 2 p 2 g(x 2 ) x 3 p 3 g(x 3 ) ………... x n p n g(x n ) Then, one at a time, you weight the value of the function by its corresponding probability. EXPECTED VALUE OF A FUNCTION OF A RANDOM VARIABLE

© Christopher Dougherty 1999–2006 x i p i g(x i ) g(x i ) p i x 1 p 1 g(x 1 )g(x 1 ) p 1 x 2 p 2 g(x 2 ) g(x 2 ) p 2 x 3 p 3 g(x 3 ) g(x 3 ) p 3 ………...……... x n p n g(x n ) g(x n ) p n You do this individually for each possible value of X. EXPECTED VALUE OF A FUNCTION OF A RANDOM VARIABLE

© Christopher Dougherty 1999–2006 x i p i g(x i ) g(x i ) p i x 1 p 1 g(x 1 )g(x 1 ) p 1 x 2 p 2 g(x 2 ) g(x 2 ) p 2 x 3 p 3 g(x 3 ) g(x 3 ) p 3 ………...……... x n p n g(x n ) g(x n ) p n   g(x i ) p i The sum of the weighted values is the expected value of the function of X. EXPECTED VALUE OF A FUNCTION OF A RANDOM VARIABLE

© Christopher Dougherty 1999–2006 x i p i g(x i ) g(x i ) p i x i p i x 1 p 1 g(x 1 )g(x 1 ) p 1 21/36 x 2 p 2 g(x 2 ) g(x 2 ) p 2 32/36 x 3 p 3 g(x 3 ) g(x 3 ) p 3 43/36 ………...……... 54/36 ………...……... 65/36 ………...……... 76/36 ………...……... 85/36 ………...……... 94/36 ………...…… /36 ………...…… /36 x n p n g(x n ) g(x n ) p n 121/36   g(x i ) p i The process will be illustrated for X 2, where X is the random variable defined in the first sequence. The 11 possible values of X and the corresponding probabilities are listed. EXPECTED VALUE OF A FUNCTION OF A RANDOM VARIABLE

© Christopher Dougherty 1999–2006 x i p i g(x i ) g(x i ) p i x i p i x i 2 x 1 p 1 g(x 1 )g(x 1 ) p 1 21/364 x 2 p 2 g(x 2 ) g(x 2 ) p 2 32/369 x 3 p 3 g(x 3 ) g(x 3 ) p 3 43/3616 ………...……... 54/3625 ………...……... 65/3636 ………...……... 76/3649 ………...……... 85/3664 ………...……... 94/3681 ………...…… /36100 ………...…… /36121 x n p n g(x n ) g(x n ) p n 121/36144   g(x i ) p i First you calculate the possible values of X 2. EXPECTED VALUE OF A FUNCTION OF A RANDOM VARIABLE

© Christopher Dougherty 1999–2006 x i p i g(x i ) g(x i ) p i x i p i x i 2 x i 2 p i x 1 p 1 g(x 1 )g(x 1 ) p 1 21/ x 2 p 2 g(x 2 ) g(x 2 ) p 2 32/369 x 3 p 3 g(x 3 ) g(x 3 ) p 3 43/3616 ………...……... 54/3625 ………...……... 65/3636 ………...……... 76/3649 ………...……... 85/3664 ………...……... 94/3681 ………...…… /36100 ………...…… /36121 x n p n g(x n ) g(x n ) p n 121/36144   g(x i ) p i The first value is 4, which arises when X is equal to 2. The probability of X being equal to 2 is 1/36, so the weighted function is 4/36, which we shall write in decimal form as EXPECTED VALUE OF A FUNCTION OF A RANDOM VARIABLE

© Christopher Dougherty 1999–2006 x i p i g(x i ) g(x i ) p i x i p i x i 2 x i 2 p i x 1 p 1 g(x 1 )g(x 1 ) p 1 21/ x 2 p 2 g(x 2 ) g(x 2 ) p 2 32/ x 3 p 3 g(x 3 ) g(x 3 ) p 3 43/ ………...……... 54/ ………...……... 65/ ………...……... 76/ ………...……... 85/ ………...……... 94/ ………...…… / ………...…… / x n p n g(x n ) g(x n ) p n 121/   g(x i ) p i Similarly for all the other possible values of X. EXPECTED VALUE OF A FUNCTION OF A RANDOM VARIABLE

© Christopher Dougherty 1999–2006 x i p i g(x i ) g(x i ) p i x i p i x i 2 x i 2 p i x 1 p 1 g(x 1 )g(x 1 ) p 1 21/ x 2 p 2 g(x 2 ) g(x 2 ) p 2 32/ x 3 p 3 g(x 3 ) g(x 3 ) p 3 43/ ………...……... 54/ ………...……... 65/ ………...……... 76/ ………...……... 85/ ………...……... 94/ ………...…… / ………...…… / x n p n g(x n ) g(x n ) p n 121/   g(x i ) p i The expected value of X 2 is the sum of its weighted values in the final column. It is equal to It is the average value of the figures in the previous column, taking the differing probabilities into account. EXPECTED VALUE OF A FUNCTION OF A RANDOM VARIABLE

© Christopher Dougherty 1999–2006 x i p i g(x i ) g(x i ) p i x i p i x i 2 x i 2 p i x 1 p 1 g(x 1 )g(x 1 ) p 1 21/ x 2 p 2 g(x 2 ) g(x 2 ) p 2 32/ x 3 p 3 g(x 3 ) g(x 3 ) p 3 43/ ………...……... 54/ ………...……... 65/ ………...……... 76/ ………...……... 85/ ………...……... 94/ ………...…… / ………...…… / x n p n g(x n ) g(x n ) p n 121/   g(x i ) p i Note that E(X 2 ) is not the same thing as E(X), squared. In the previous sequence we saw that E(X) for this example was 7. Its square is 49.

© Christopher Dougherty 1999–2006 Population variance of X: The previous sequence defined the expected value of a function of a random variable X. There is only one function that is of much interest to us, at least initially: the squared deviation from the population mean. The expected value of the squared deviation is known as the population variance of X. It is a measure of the dispersion of the distribution of X about its population mean. POPULATION VARIANCE OF A DISCRETE RANDOM VARIABLE

© Christopher Dougherty 1999–2006 We will calculate the population variance of the random variable X defined in the first sequence. We start as usual by listing the possible values of X and the corresponding probabilities. POPULATION VARIANCE OF A DISCRETE RANDOM VARIABLE x i p i x i –  (x i –  ) 2 (x i –  ) 2 p i 21/36– /36– /36– /36– /36– / / / / / /

© Christopher Dougherty 1999–2006 Next we need a column giving the deviations of the possible values of X about its population mean. In the second sequence we saw that the population mean of X was 7. POPULATION VARIANCE OF A DISCRETE RANDOM VARIABLE x i p i x i –  (x i –  ) 2 (x i –  ) 2 p i 21/36– /36– /36– /36– /36– / / / / / /

© Christopher Dougherty 1999–2006 When X is equal to 2, the deviation is –5. POPULATION VARIANCE OF A DISCRETE RANDOM VARIABLE x i p i x i –  (x i –  ) 2 (x i –  ) 2 p i 21/36– /36– /36– /36– /36– / / / / / /

© Christopher Dougherty 1999–2006 POPULATION VARIANCE OF A DISCRETE RANDOM VARIABLE x i p i x i –  (x i –  ) 2 (x i –  ) 2 p i 21/36– /36– /36– /36– /36– / / / / / /

© Christopher Dougherty 1999–2006 Next we need a column giving the squared deviations. When X is equal to 2, the squared deviation is 25. POPULATION VARIANCE OF A DISCRETE RANDOM VARIABLE x i p i x i –  (x i –  ) 2 (x i –  ) 2 p i 21/36– /36– /36– /36– /36– / / / / / /

© Christopher Dougherty 1999–2006 Similarly for the other values of X. POPULATION VARIANCE OF A DISCRETE RANDOM VARIABLE x i p i x i –  (x i –  ) 2 (x i –  ) 2 p i 21/36– /36– /36– /36– /36– / / / / / /

© Christopher Dougherty 1999–2006 POPULATION VARIANCE OF A DISCRETE RANDOM VARIABLE x i p i x i –  (x i –  ) 2 (x i –  ) 2 p i 21/36– /36– /36– /36– /36– / / / / / / Now we start weighting the squared deviations by the corresponding probabilities. What do you think the weighted average will be? Have a guess.

© Christopher Dougherty 1999–2006 A reason for making an initial guess is that it may help you to identify an arithmetical error, if you make one. If the initial guess and the outcome are very different, that is a warning. POPULATION VARIANCE OF A DISCRETE RANDOM VARIABLE x i p i x i –  (x i –  ) 2 (x i –  ) 2 p i 21/36– /36– /36– /36– /36– / / / / / /

© Christopher Dougherty 1999–2006 We calculate all the weighted squared deviations. POPULATION VARIANCE OF A DISCRETE RANDOM VARIABLE x i p i x i –  (x i –  ) 2 (x i –  ) 2 p i 21/36– /36– /36– /36– /36– / / / / / /

© Christopher Dougherty 1999–2006 x i p i x i –  (x i –  ) 2 (x i –  ) 2 p i 21/36– /36– /36– /36– /36– / / / / / / The sum is the population variance of X. POPULATION VARIANCE OF A DISCRETE RANDOM VARIABLE

© Christopher Dougherty 1999–2006 Population variance of X In equations, the population variance of X is usually written  X 2,  being the Greek s. POPULATION VARIANCE OF A DISCRETE RANDOM VARIABLE

© Christopher Dougherty 1999–2006 Standard deviation of X The standard deviation of X is the square root of its population variance. Usually written  x, it is an alternative measure of dispersion. It has the same units as X. POPULATION VARIANCE OF A DISCRETE RANDOM VARIABLE

Properties of continuous random variables

© Christopher Dougherty 1999–2006 CONTINUOUS RANDOM VARIABLES A discrete random variable is one that can take only a finite set of values. The sum of the numbers when two dice are thrown is an example. Each value has associated with it a finite probability, which you can think of as a ‘packet’ of probability. The packets sum to unity because the variable must take one of the values. 6 __ 36 5 __ 36 4 __ 36 3 __ 36 2 __ 36 2 __ 36 3 __ 36 5 __ 36 4 __ 36 probability X

© Christopher Dougherty 1999– X 65 CONTINUOUS RANDOM VARIABLES height However, most random variables encountered in econometrics are continuous. They can take any one of an infinite set of values defined over a range (or possibly, ranges). As a simple example, take the temperature in a room. We will assume that it can be anywhere from 55 to 75 degrees Fahrenheit with equal probability within the range.

© Christopher Dougherty 1999– X 65 CONTINUOUS RANDOM VARIABLES height In the case of a continuous random variable, the probability of it being equal to a given finite value (for example, temperature equal to ) is always infinitesimal. For this reason, you can only talk about the probability of a continuous random variable lying between two given values. The probability is represented graphically as an area.

© Christopher Dougherty 1999– X 65 CONTINUOUS RANDOM VARIABLES 7 height 56 For example, you could measure the probability of the temperature being between 55 and 56, both measured exactly. Given that the temperature lies anywhere between 55 and 75 with equal probability, the probability of it lying between 55 and 56 must be 0.05.

© Christopher Dougherty 1999– X 65 CONTINUOUS RANDOM VARIABLES 57 height Similarly, the probability of the temperature lying between 56 and 57 is

© Christopher Dougherty 1999– X 65 CONTINUOUS RANDOM VARIABLES height And similarly for all the other one-degree intervals within the range. 0.05

© Christopher Dougherty 1999– X CONTINUOUS RANDOM VARIABLES height The probability per unit interval is 0.05 and accordingly the area of the rectangle representing the probability of the temperature lying in any given unit interval is The probability per unit interval is called the probability density and it is equal to the height of the unit-interval rectangle.

© Christopher Dougherty 1999– X CONTINUOUS RANDOM VARIABLES f(X) = 0.05 for 55 X 75 f(X) = 0 for X Mathematically, the probability density is written as a function of the variable, for example f(X). In this example, f(X) is 0.05 for 55 < X < 75 and it is zero elsewhere. height

© Christopher Dougherty 1999– X CONTINUOUS RANDOM VARIABLES probability density f(X)f(X) f(X) = 0.05 for 55 X 75 f(X) = 0 for X The vertical axis is given the label probability density, rather than height. f(X) is known as the probability density function and is shown graphically in the diagram as the thick black line.

© Christopher Dougherty 1999– X CONTINUOUS RANDOM VARIABLES probability density f(X)f(X) f(X) = 0.05 for 55 X 75 f(X) = 0 for X 75 Suppose that you wish to calculate the probability of the temperature lying between 65 and 70 degrees. To do this, you should calculate the area under the probability density function between 65 and 70. Typically you have to use the integral calculus to work out the area under a curve, but in this very simple example all you have to do is calculate the area of a rectangle.

© Christopher Dougherty 1999– X CONTINUOUS RANDOM VARIABLES The height of the rectangle is 0.05 and its width is 5, so its area is probability density f(X)f(X) f(X) = 0.05 for 55 X 75 f(X) = 0 for X 75