Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.

Slides:



Advertisements
Similar presentations
Hypothesis testing Another judgment method of sampling data.
Advertisements

CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Sampling: Final and Initial Sample Size Determination
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
DATA ANALYSIS Module Code: CA660 Lecture Block 5.
Chap 8: Estimation of parameters & Fitting of Probability Distributions Section 6.1: INTRODUCTION Unknown parameter(s) values must be estimated before.
DATA ANALYSIS Module Code: CA660 Lecture Block 6: Alternative estimation methods and their implementation.
Maximum likelihood (ML) and likelihood ratio (LR) test
CS 8751 ML & KDDEvaluating Hypotheses1 Sample error, true error Confidence intervals for observed hypothesis error Estimators Binomial distribution, Normal.
Statistical Inference Chapter 12/13. COMP 5340/6340 Statistical Inference2 Statistical Inference Given a sample of observations from a population, the.
Maximum likelihood (ML)
Chapter 7 Sampling and Sampling Distributions
9-1 Hypothesis Testing Statistical Hypotheses Statistical hypothesis testing and confidence interval estimation of parameters are the fundamental.
Maximum likelihood (ML) and likelihood ratio (LR) test
1/55 EF 507 QUANTITATIVE METHODS FOR ECONOMICS AND FINANCE FALL 2008 Chapter 10 Hypothesis Testing.
Topic 2: Statistical Concepts and Market Returns
Evaluating Hypotheses
Parametric Inference.
Lec 6, Ch.5, pp90-105: Statistics (Objectives) Understand basic principles of statistics through reading these pages, especially… Know well about the normal.
DATA ANALYSIS Module Code: CA660 Lecture Block 5.
1 STATISTICAL INFERENCE PART I EXPONENTIAL FAMILY & POINT ESTIMATION.
Aaker, Kumar, Day Seventh Edition Instructor’s Presentation Slides
Inferences About Process Quality
Lehrstuhl für Informatik 2 Gabriella Kókai: Maschine Learning 1 Evaluating Hypotheses.
BCOR 1020 Business Statistics Lecture 18 – March 20, 2008.
BCOR 1020 Business Statistics
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
5-3 Inference on the Means of Two Populations, Variances Unknown
Maximum likelihood (ML)
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
Chapter 9 Title and Outline 1 9 Tests of Hypotheses for a Single Sample 9-1 Hypothesis Testing Statistical Hypotheses Tests of Statistical.
Aaker, Kumar, Day Ninth Edition Instructor’s Presentation Slides
Lecture 5: Segregation Analysis I Date: 9/10/02  Counting number of genotypes, mating types  Segregation analysis: dominant, codominant, estimating segregation.
Model Inference and Averaging
Lecture 14 Sections 7.1 – 7.2 Objectives:
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Random Sampling, Point Estimation and Maximum Likelihood.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
1 MAXIMUM LIKELIHOOD ESTIMATION Recall general discussion on Estimation, definition of Likelihood function for a vector of parameters  and set of values.
9-1 Hypothesis Testing Statistical Hypotheses Definition Statistical hypothesis testing and confidence interval estimation of parameters are.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
Chapter 7 Point Estimation
4 Hypothesis & Testing. CHAPTER OUTLINE 4-1 STATISTICAL INFERENCE 4-2 POINT ESTIMATION 4-3 HYPOTHESIS TESTING Statistical Hypotheses Testing.
1 Lecture 16: Point Estimation Concepts and Methods Devore, Ch
Lecture 12: Linkage Analysis V Date: 10/03/02  Least squares  An EM algorithm  Simulated distribution  Marker coverage and density.
Lecture 3: Statistics Review I Date: 9/3/02  Distributions  Likelihood  Hypothesis tests.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Lecture 22: Quantitative Traits II
Lecture 23: Quantitative Traits III Date: 11/12/02  Single locus backcross regression  Single locus backcross likelihood  F2 – regression, likelihood,
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University.
Chapter 8 Estimation ©. Estimator and Estimate estimator estimate An estimator of a population parameter is a random variable that depends on the sample.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Hypothesis Testing. Statistical Inference – dealing with parameter and model uncertainty  Confidence Intervals (credible intervals)  Hypothesis Tests.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 7 Inferences Concerning Means.
Estimating standard error using bootstrap
Chapter 4. Inference about Process Quality
STATISTICAL INFERENCE PART I POINT ESTIMATION
CONCEPTS OF ESTIMATION
9 Tests of Hypotheses for a Single Sample CHAPTER OUTLINE
Simple Linear Regression
Inference on the Mean of a Population -Variance Known
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Presentation transcript:

Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties of estimators

Types of Errors  False positive (Type I): Probability (  ) that H 0 rejected when it is true.  False negative (Type II): Probability (  ) of accepting H 0 when it is false. Accept H 0 Reject H 0 H 0 true 1-  Type I =  H 0 false Type II =  power = 1- 

Example: Error H 0 : Central Chi-Square H A : Non-Central Chi-Square with non-centrality parameter E(G)   1-  1- 

Power of a Test  Definition: The statistical power of a test is defined as 1-   The power is only defined when H A is defined, the experimental conditions (e.g. sample size) are known and the significance level  has been selected.  Example: calculate sample size needed to obtain particular linkage detection power.

Estimation  Hypothesis testing allows us to make qualitative conclusions regarding the suitability or not of a statement (H 0 ).  Often we want to make quantitative inference, e.g. an actual estimate of the recombination fraction, not just evidence that genes are linked.

Estimation  Definition: Point estimation is the process of estimating a specific value of the parameter  based on observed data X 1, X 2,…,X n.  Definition: Interval estimation is the process of estimating upper and lower limits within which the unknown parameter occurs with certain probability.

 Definition: The maximum likelihood estimator is the which maximizes the likelihood function.  The MLE is obtained by analytically solving the score, S=0 grid search Newton-Raphson iteration Expectation and Maximization (EM) algorithm Maximum Likelihood Estimation

MLE: Grid Search  Plot likelihood L or log likelihood l vs. parameter throughout the parameter space.  Obtain MLE by visual inspection or search algorithm.

MLE: Grid Search Algorithm 1.Initially estimate  0. Pick a step size . 2.At step n, evaluate L (or l) on both sides of  n. 3.Choose  n+1 =  n +  if L is increasing to the right, else choose  n+1 =  n - . 4.Repeat steps 2 and 3 until no longer advance. Choose smaller , and repeat steps 2, 3, and 4 until desired accuracy met.

MLE: Grid Search Problems  Multiple peaks result in a failure to find the global maximum likelihood.  Solving for multiple simultaneous parameters gets computationally intensive and difficult to interpret visually when there are more than two parameters.

Example: 2-Locus with Non- Penetrant Allele  Let  be the recombinant fraction between marker A and gene of interest B.  Let  be the probability that the allele of the gene of interest (f) fails to be detected at the phenotype level (i.e. 1-  is the penetrance).  Cross +F/+F  –f/–f.  Score gametes of an F1 +F/–f individual for +/- phenotype and P/p phenotype, where P means F or non-penetrant f and p means penetrant f. B(F or f) A(+ or -)

Experimental Data  P(+F gamete) = P(–f gamete) = 0.5(1-  )  P(–F gamete) = P(+f gamete) = 0.5   P(+P gamete) = 0.5(1-  )   P(–P gamete) = 0.5  + 0.5(1-  )   P(+p gamete) = 0.5  (1-  )  P(–p gamete) = 0.5(1-  ) (1-  )  Observe n +P, n -P, n +p, n -p.

Experimental Data: Log Likelihood

Experimental Data: Grid Search

Newton-Raphson: One Parameter  Let S(  ) be the score.  Then the MLE is obtained through the equation  Taylor expansion of S(  ) for  n near the MLE, gives

Newton-Raphson: Analysis  NR fits a parabola to the likelihood function at the point of the current parameter estimate. Obtain a new parameter estimate at the maximum of the parabola.  NR may fail when there are multiple peaks.  NR may fail when the information is zero (when the estimate is at the extremes).  Recommendations: Use multiple starting initial values. Bound new estimates.

Newton-Raphson: Multiple Parameters N is the total sample size. S(  m ) is the score vector evaluated at  m. I -1 (  m ) is the inverse information matrix evaluated at  m.

EM Algorithm: Incomplete Data  The notion of incomplete data: ABabAbaB ABAB/ABAB/abAB/AbAB/aB abab/ABab/abab/Abab/aB AbAb/ABAb/abAb/AbAb/aB aBaB/ABaB/abaB/AbaB/aB

EM Algorithm: Example of Incomplete Data  +P gamete may result from nonrecombinant +F or from recombinant, non-penetrant +f.  +p gamete can only result from penetrant, nonrecombinant +f.  – P gamete can result from recombinant – F or from nonrecombinant, non-penetrant – f gene.  –p gamete can result only from nonrecombinant –f.

EM Algorithm  Make an initial guess  0.  Expectation step: Pretend that  n for iteration n is true. Estimate the complete data. This usually request distribution of complete data conditional on the observed data. For example: P(recombinant|observed phenotype).  Maximization step: Compute the maximum likelihood estimate for step n.  Repeat E & M steps until likelihood converges.

Example: E Step  E Step:

Example: M Step  M Step:

Moment Estimation  Obtain equations for the population moments in terms of the parameters to estimate.  Substitute the sample moments and solve for the parameters.  For example: binomial distribution m 1 = np

Example: Moment Estimation

Moment Estimation: Problems  Large sample properties of ML estimators are usually better than those for the corresponding moment estimators.  Sometimes solution of moments equations are not unique.

Least Squares Estimation

Variance of an Estimator Suppose k independent estimates are available for  : Suppose you have a large sample, then the variance of the MLE is approximately: Cramer-Rao lower bound for variance Empirical estimates using resampling techniques.

Variance: Linear Estimator

Variance: General Function f(  ,  , …,   )

Bias  The mean square estimator (MSE) is defined as bias If an estimator is unbiased, the MSE and variance are the same.

Estimating Bias  Bootstrap: bootstrap estimator for bootstrap trial i original estimate

Confidence Interval  Because of sampling error, the estimate is not exactly the true parameter value .  A confidence interval is symmetric if  A confidence interval is non-symmetric if  A confidence interval is one-sided if

Confidence Interval: Normal Approximation I  Need pivotal quantity, i.e. a quantity that depends on the data and the parameters but whose distribution does not.  If the estimate is unbiased and normally distributed with variance, then the pivotal quantity is

Confidence Interval: Normal Approximation II  The MLE is asymptotically normally distributed.

Confidence Interval: Nonparametric Approximation percentile method

Bootstrap Example  Generate a multinomial random variable with the given proportions b times and generate a bootstrap dataset. Estimate parameters  and. +P+p+p –P–P–p–p Count Proportion

Confidence Interval: Likelihood Approach  Let L max be the maximum likelihood for a given model. Find the parameter values  L and  U such that log L max – log L(   ) = 2 Then (  L,  U ) serves as a confidence interval.

LOD Score Support  The LOD score support for a confidence interval is log 10 L max –log 10 L  where L is the likelihood at the limit values of the parameter.  In practice, you plot the LOD score support for various values of the parameter and choose the upper and lower bounds such that the LOD score support is 1.

Choosing Good Confidence Intervals  The actual coverage probability should be close to the confidence coefficient.  Should be biologically relevant. For example, the following is not:  (0.1,0.6)

Good Estimator: Consistent  An estimator is mean squared error consistent if the MSE approaches zero as the sample size approaches infinity.  An estimator is simple consistent if

Good Estimator: Unbiased  An unbiased estimator is usually better than a biased one, but this may not always be true. If the variance is larger, what have we gained?  There are bootstrap techniques for obtaining a bias-corrected estimate. These are computationally more intensive than bootstrap, but sometimes worth it.

Good Estimator: Asymptotically Normal  If the pivotal quantity  is normal with mean 0 and variance 1 as the sample size goes to infinity, it can be a very convenient property of the estimator.`

Good Estimator: Confidence Interval  A good estimator should have a good way to obtain an confidence interval. MLE are good in this way if the sample size is large enough.

Sample Size for Power   1-  1-  Need E(G)> E(G)=nE(G unit )

Sample Size for Target Confidence Interval  Confidence interval by normal approximation is  The bigger the range, the less precise the confidence interval.  Suppose we wish to have

Sample Size for Target Confidence Interval II  Then,

Summary  Distributions  Likelihood and Maximum Likelihood Estimation  Hypothesis Tests  Confidence Intervals  Comparison of estimators  Sample size calculations