Lab 3b: Distribution of the mean

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

The Maximum Likelihood Method
THE CENTRAL LIMIT THEOREM
Welcome to PHYS 225a Lab Introduction, class rules, error analysis Julia Velkovska.
Evaluating Classifiers
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Sampling: Final and Initial Sample Size Determination
Introduction to Statistics
ELEC 303 – Random Signals Lecture 18 – Statistics, Confidence Intervals Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 10, 2009.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
Chapter 6 Introduction to Sampling Distributions
Chapter 7 Sampling and Sampling Distributions
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 6 Introduction to Sampling Distributions.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Evaluating Hypotheses
Lehrstuhl für Informatik 2 Gabriella Kókai: Maschine Learning 1 Evaluating Hypotheses.
Statistical inference Population - collection of all subjects or objects of interest (not necessarily people) Sample - subset of the population used to.
Chapter 7 ~ Sample Variability
QUIZ CHAPTER Seven Psy302 Quantitative Methods. 1. A distribution of all sample means or sample variances that could be obtained in samples of a given.
Standard error of estimate & Confidence interval.
Sampling Distributions  A statistic is random in value … it changes from sample to sample.  The probability distribution of a statistic is called a sampling.
Review of normal distribution. Exercise Solution.
Physics 114: Lecture 15 Probability Tests & Linear Fitting Dale E. Gary NJIT Physics Department.
Sampling Theory Determining the distribution of Sample statistics.
Albert Morlan Caitrin Carroll Savannah Andrews Richard Saney.
AP Statistics Chapter 9 Notes.
880.P20 Winter 2006 Richard Kass 1 Maximum Likelihood Method (MLM) Does this procedure make sense? The MLM answers this question and provides a method.
Random Sampling, Point Estimation and Maximum Likelihood.
Copyright ©2011 Nelson Education Limited The Normal Probability Distribution CHAPTER 6.
R. Kass/W03P416/Lecture 7 1 Lecture 7 Some Advanced Topics using Propagation of Errors and Least Squares Fitting Error on the mean (review from Lecture.
Section 8.1 Estimating  When  is Known In this section, we develop techniques for estimating the population mean μ using sample data. We assume that.
Chapter 6 Lecture 3 Sections: 6.4 – 6.5.
Chapter 7: Sample Variability Empirical Distribution of Sample Means.
Lecture 2 Review Probabilities Probability Distributions Normal probability distributions Sampling distributions and estimation.
1 Chapter 7 Sampling Distributions. 2 Chapter Outline  Selecting A Sample  Point Estimation  Introduction to Sampling Distributions  Sampling Distribution.
8 Sampling Distribution of the Mean Chapter8 p Sampling Distributions Population mean and standard deviation,  and   unknown Maximal Likelihood.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
1 2 nd Pre-Lab Quiz 3 rd Pre-Lab Quiz 4 th Pre-Lab Quiz.
LECTURE 3: ANALYSIS OF EXPERIMENTAL DATA
HYPOTHESIS TESTING Distributions(continued); Maximum Likelihood; Parametric hypothesis tests (chi-squared goodness of fit, t-test, F-test) LECTURE 2 Supplementary.
7 sum of RVs. 7-1: variance of Z Find the variance of Z = X+Y by using Var(X), Var(Y), and Cov(X,Y)
Confidence Interval & Unbiased Estimator Review and Foreword.
Summarizing Risk Analysis Results To quantify the risk of an output variable, 3 properties must be estimated: A measure of central tendency (e.g. µ ) A.
Introduction to Inference Sampling Distributions.
Sampling Distributions Chapter 18. Sampling Distributions If we could take every possible sample of the same size (n) from a population, we would create.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Sampling Theory Determining the distribution of Sample statistics.
Learning Theory Reza Shadmehr Distribution of the ML estimates of model parameters Signal dependent noise models.
CHAPTER 4 ESTIMATES OF MEAN AND ERRORS. 4.1 METHOD OF LEAST SQUARES I n Chapter 2 we defined the mean  of the parent distribution and noted that the.
Richard Kass/F02P416 Lecture 6 1 Lecture 6 Chi Square Distribution (  2 ) and Least Squares Fitting Chi Square Distribution (  2 ) (See Taylor Ch 8,
The accuracy of averages We learned how to make inference from the sample to the population: Counting the percentages. Here we begin to learn how to make.
Sums of Random Variables and Long-Term Averages Sums of R.V. ‘s S n = X 1 + X X n of course.
R. Kass/Sp07P416/Lecture 71 More on Least Squares Fit (LSQF) In Lec 5, we discussed how we can fit our data points to a linear function (straight line)
Stats 242.3(02) Statistical Theory and Methodology.
Confidence Intervals Cont.
The Maximum Likelihood Method
Physics 114: Lecture 13 Probability Tests & Linear Fitting
Chapter 7 Review.
Introduction, class rules, error analysis Julia Velkovska
The Maximum Likelihood Method
Sample Mean Distributions
Combining Random Variables
The Maximum Likelihood Method
Statistical Methods For Engineers
CONCEPTS OF ESTIMATION
Sampling Distribution Models
STATISTICAL INFERENCE PART IV
Stat Lab 9.
Quantitative Methods Varsha Varde.
Presentation transcript:

Lab 3b: Distribution of the mean Outline Distribution of the mean: Normal (Gaussian) Central Limit Theorem “Justification” of the mean as the best estimate “Justification” of error propagation formula Propagation of errors Standard deviation of the mean 640:104 - Elementary Combinatorics and Probability 960:211-212. Statistics I, II

Finite number of repeated measurements The average value is the best estimation of true mean value, while the square of standard deviation is the best estimation of true variance. The same set of data also provides the best estimation of the variance of the mean.

The Central Limit Theorem The mean is a sum of N independent random variables: The PDF of the mean is the PDF of the sum of N independent random variables. In probability theory, the central limit theorem (CLT) states that, given conditions, the mean of a sufficiently large number of independent random variables, each with finite mean and variance, will be approximately normally distributed. *In practice, N just needs to be large enough.

Probability Distribution Function of the mean An example of CLT For this example N>4 is “enough”.

More examples of CLT http://www.statisticalengineering.com/central_limit_theorem.html

The significant consequence of CLT Almost any repeated measurements of independent random variables can be effectively described by Normal (Gaussian) distribution. E.g. N measurements xi (i=1, 2, …, N) of random variable x can be divided to m subsets of measurements {nj}, j=1,2,…,m. Define the means of each subset as new measurements of new random variable y. If nj are large enough, the PDF of yj is approximately a Normal (Gaussian) one.

“Justification” of the mean as the best estimate Assuming all the measured values follow a normal distribution GX,(x) with unknown parameters X and . The probability of obtaining value x1 in a small interval dx1 is: constant Similarly, the probability of obtaining value xi (i=0, …, N) in a small interval dxi is:

Justification of the mean as best estimate (cont.) The probability of N measurements with values: x1, x2, … xN: Here x1, x2, … xN are known results because they are measured values. On the other hand, X and  are unknown parameters. We want to find out the best estimate of them from ProbX, (x1, x2, … xN ). One of the reasonable procedures is the principle of maximum likelihood, i.e. maximizing ProbX, (x1, x2, … xN ) respect to X minimizing

Best estimate of variance Similarly, we could maximize ProbX, (x1, x2, … xN ) respect to  to get the best estimate of : However, we don’t really know X. In practice, we have to replace X with the mean value, which reduces the accuracy. To account for this factor, we have to replace N with N-1.

Justification of error propagation formula (I) Measured quantity plus a constant: Measured quantity multiplies a fixed #: Sum of two measured quantities: Assuming X=Y=0,

Justification of error propagation formula (II) Sum of two measured quantities:

Justification of error propagation formula (III) constant

Justification of error propagation formula (IV)

Standard deviation of the mean

Lab 3b: test the CLT One run of decay rate (R=N/Δt) measurement Sampling rate: 0.5 second/sample (Δt) Run time: 30 seconds (# of measurements: n=60) calculate , plot histogram. Repeat the 30-sec run 20 times (a set of runs) Remember to save data Plot histograms of 21 means with proper bin widths Make sure you answer all the questions Sample size (n=T/ Δt) dependence Keep the same configuration, vary run time * “Origin” (OriginLab®) is more convenient than Matlab. http://www.physics.rutgers.edu/rcem/~ahogan/326.html