A) Transformation method (for continuous distributions) U(0,1) : uniform distribution f(x) : arbitrary distribution f(x) dx = U(0,1)(u) du When inverse.

Slides:



Advertisements
Similar presentations
Point Estimation Notes of STAT 6205 by Dr. Fan.
Advertisements

Probability and Statistics Basic concepts II (from a physicist point of view) Benoit CLEMENT – Université J. Fourier / LPSC
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
Sampling: Final and Initial Sample Size Determination
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
Parameter Estimation: Maximum Likelihood Estimation Chapter 3 (Duda et al.) – Sections CS479/679 Pattern Recognition Dr. George Bebis.
Maximum likelihood (ML) and likelihood ratio (LR) test
The Mean Square Error (MSE):. Now, Examples: 1) 2)
Hypothesis testing Some general concepts: Null hypothesisH 0 A statement we “wish” to refute Alternative hypotesisH 1 The whole or part of the complement.
Maximum likelihood Conditional distribution and likelihood Maximum likelihood estimations Information in the data and likelihood Observed and Fisher’s.
G. Cowan Lectures on Statistical Data Analysis Lecture 2 page 1 Statistical Data Analysis: Lecture 2 1Probability, Bayes’ theorem 2Random variables and.
Chapter 6 Introduction to Sampling Distributions
Statistical Inference Chapter 12/13. COMP 5340/6340 Statistical Inference2 Statistical Inference Given a sample of observations from a population, the.
Part 2b Parameter Estimation CSE717, FALL 2008 CUBS, Univ at Buffalo.
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 6 Introduction to Sampling Distributions.
Maximum likelihood (ML) and likelihood ratio (LR) test
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Evaluating Hypotheses
Generating Continuous Random Variables some. Quasi-random numbers So far, we learned about pseudo-random sequences and a common method for generating.
2. Point and interval estimation Introduction Properties of estimators Finite sample size Asymptotic properties Construction methods Method of moments.
Probability and Statistics Review
Statistical Background
7. Least squares 7.1 Method of least squares K. Desch – Statistical methods of data analysis SS10 Another important method to estimate parameters Connection.
5. Estimation 5.3 Estimation of the mean K. Desch – Statistical methods of data analysis SS10 Is an efficient estimator for μ ?  depends on the distribution.
K. Desch – Statistical methods of data analysis SS10 3. Distributions 3.1 Binomial distribution Error of an efficiency Estimator for the efficiency (when.
STATISTICAL INFERENCE PART II SOME PROPERTIES OF ESTIMATORS
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Rao-Cramer-Frechet (RCF) bound of minimum variance (w/o proof) Variance of an estimator of single parameter is limited as: is called “efficient” when the.
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
Maximum likelihood (ML)
The Monte Carlo Method: an Introduction Detlev Reiter Research Centre Jülich (FZJ) D Jülich
Monte Carlo Integration Digital Image Synthesis Yung-Yu Chuang with slides by Pat Hanrahan and Torsten Moller.
Chapter 5 Sampling and Statistics Math 6203 Fall 2009 Instructor: Ayona Chatterjee.
Simulation Output Analysis
Analysis of Monte Carlo Integration Fall 2012 By Yaohang Li, Ph.D.
 1  Outline  stages and topics in simulation  generation of random variates.
880.P20 Winter 2006 Richard Kass 1 Confidence Intervals and Upper Limits Confidence intervals (CI) are related to confidence limits (CL). To calculate.
Continuous Probability Distributions  Continuous Random Variable  A random variable whose space (set of possible values) is an entire interval of numbers.
Random Sampling, Point Estimation and Maximum Likelihood.
Monte Carlo I Previous lecture Analytical illumination formula This lecture Numerical evaluation of illumination Review random variables and probability.
1 G Lect 3b G Lecture 3b Why are means and variances so useful? Recap of random variables and expectations with examples Further consideration.
Chapter 7 Point Estimation
Monte Carlo Methods So far we have discussed Monte Carlo methods based on a uniform distribution of random numbers on the interval [0,1] p(x) = 1 0  x.
Sample variance and sample error We learned recently how to determine the sample variance using the sample mean. How do we translate this to an unbiased.
1 Lecture 16: Point Estimation Concepts and Methods Devore, Ch
Chapter 5 Parameter estimation. What is sample inference? Distinguish between managerial & financial accounting. Understand how managers can use accounting.
Functions of random variables Sometimes what we can measure is not what we are interested in! Example: mass of binary-star system: We want M but can only.
Chapter 2 Statistical Background. 2.3 Random Variables and Probability Distributions A variable X is said to be a random variable (rv) if for every real.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
1 Standard error Estimated standard error,s,. 2 Example 1 While measuring the thermal conductivity of Armco iron, using a temperature of 100F and a power.
Confidence Interval & Unbiased Estimator Review and Foreword.
Sampling and estimation Petter Mostad
Brief Review Probability and Statistics. Probability distributions Continuous distributions.
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
Statistical Estimation Vasileios Hatzivassiloglou University of Texas at Dallas.
Analysis of Experimental Data; Introduction
ES 07 These slides can be found at optimized for Windows)
Week 21 Order Statistics The order statistics of a set of random variables X 1, X 2,…, X n are the same random variables arranged in increasing order.
Gil McVean, Department of Statistics Thursday February 12 th 2009 Monte Carlo simulation.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate its.
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Monte Carlo Integration Digital Image Synthesis Yung-Yu Chuang 12/3/2008 with slides by Pat Hanrahan and Torsten Moller.
Confidence Intervals and Sample Size
The Maximum Likelihood Method
Statistical Methods for Data Analysis Random number generators
Statistical Inference
Applied Statistics and Probability for Engineers
Maximum Likelihood Estimation (MLE)
Presentation transcript:

a) Transformation method (for continuous distributions) U(0,1) : uniform distribution f(x) : arbitrary distribution f(x) dx = U(0,1)(u) du When inverse function of integral, F -1 (u), is known, then x = F -1 (u) distributed according to f(x) Example: Exponential distribution 4. MC Methods 4.2 Generators for arbitrary distributions K. Desch – Statistical methods of data analysis SS10

b) Transformation method (discrete distributions) 4. MC Methods 4.2 Generators for arbitrary distributions K. Desch – Statistical methods of data analysis SS10 c) Hit-or-miss method (brute force) Uniform distr. fr. 0 to c: u i Uniform distr. from x min to x max : x i when u i ≤ f(x i ) → accept x i, otherwise not - two random numbers per try - inefficient when f(x) « c - need to (conservatively) estimate c (maximum of f(x)) (can be done in “warm-up” run)

4. MC Methods 4.2 Generators for arbitrary distributions K. Desch – Statistical methods of data analysis SS10 Improvement: - search for analytical function s(x) close to f(x) - use c so that c s(x) >f(x) for all x 1.take u i in [0,1] and calculate x i = S -1 (u i ) 2.take u j in [0,c] 3.when u j s(x i ) ≤ f(x i ) accept x i, otherwise not

search for: 4. MC Methods 4.3 Monte Carlo Integration K. Desch – Statistical methods of data analysis SS10 Integration over one dimension: (E[g] = expectation value of g w.r.t. uniform distribution) Take x i uniformly distributed in [a,b] → Variance: (CLT)

4. MC Methods 4.3 Monte Carlo Integration K. Desch – Statistical methods of data analysis SS10 Alternative: hit-or-miss integration

- Variance of r(x): will be small when r is flat, so f ≈ g - The method takes care of (integrable) singularities (find f(x) with has the same singularity structure as g(x)) x i distributed as f(x) 4. MC Methods 4.3 Monte Carlo Integration K. Desch – Statistical methods of data analysis SS10 Variance-reduced methods a) importance sampling: If f(x) is a known p.d.f., which could be integrated and inverted, then: Expectation value of r(x) can be obtained with random numbers, which is distributed according to f(x):

4. MC Methods 4.3 Monte Carlo Integration K. Desch – Statistical methods of data analysis SS10 b) Control function (subtraction of an integrable analytical function) analytical MC c) Partitioning (split integration range into several more „flat“ regions)

let x be a random variable distributed according to f(x) n independent “measurements” of x, x = (x 1,…,x n ) is sample of a distribution f(x) of size n (outcome of an experiment) x = itself is a random variable with p.d.f. f sample (x) sample space: all possible values of x = (x 1,…,x n ) If all x i are independent f sample (x) = f(x 1 )f(x 2 ) … f(x n ) is the p.d.f. for x 5. Estimation 5.1 Sample space, Estimators K. Desch – Statistical methods of data analysis SS10

A central problem of (frequentist) statistics: Find the properties of f(x) when only a sample x = (x 1,…,x n ) has been measured Task: construct functions of x i to estimate the properties of f(x) (e.g. μ, σ 2, …) Often f depends on parameters θ j : f(x i ;θ j )  try to estimate the parameters θ j from measured sample x Functions of (x i ) are called a statistic. If a statistic is used to estimate parameters (μ, σ 2, θ, …), it called an estimator Notation: is an estimator for θ can be calculated; true value θ is unknown Estimation of p.d.f. parameters is also called a fit 5. Estimation 5.1 Sample Space, Estimators K. Desch – Statistical methods of data analysis SS10

in simple words: n→ ∞  θ → 2.Bias: itself is a random variable, distributed according to a p.d.f. This p.d.f. is called the sampling distribution Expectation value of the sampling distribution: (or “ “) 1 Consistency: an estimator is consistent if for each ε > 0 : 5. Estimation 5.2 Properties of Estimators K. Desch – Statistical methods of data analysis SS10 because

5. Estimation 5.2 Properties of estimators K. Desch – Statistical methods of data analysis SS10 The bias of an estimator is defined as An estimator is unbiased (or bias-free) if b=0 An estimator is asymptotically unbiased if Attentions Consistent: for large sample size Unbiased: for fixed sample size 3. Efficiency: One estimator is more efficient than another if its variance is smaller, or more precise if its mean squared error (MSE) is smaller and

5. Estimation 5.2 Properties of estimators K. Desch – Statistical methods of data analysis SS10 4. Robustness An estimator is robust if it does not strongly depend on single measurements (which might be systematically wrong) 5. Simplicity (subjective)

5. Estimation 5.3 Estimation of the mean K. Desch – Statistical methods of data analysis SS10 In principle one can construct an arbitrary number of different esitmators for the mean value of a pdf,  = E[x] Examples: mean of the sample mean of the first ten members of the sample median of the sample all have different (wanted and unwanted) properties

5. Estimation 5.3 Estimation of the mean K. Desch – Statistical methods of data analysis SS10 The mean of a sample provides an estimate of the true mean: a) is consistent: CLT: p.d.f. of approaches Gaussian with variance b) is unbiased c) Is efficient ?