DATA ANALYSIS Module Code: CA660 Lecture Block 5.

Slides:



Advertisements
Similar presentations
Hypothesis testing Another judgment method of sampling data.
Advertisements

DATA ANALYSIS Module Code: CA660 Lecture Block 5.
Chap 9: Testing Hypotheses & Assessing Goodness of Fit Section 9.1: INTRODUCTION In section 8.2, we fitted a Poisson dist’n to counts. This chapter will.
Chap 8: Estimation of parameters & Fitting of Probability Distributions Section 6.1: INTRODUCTION Unknown parameter(s) values must be estimated before.
DATA ANALYSIS Module Code: CA660 Lecture Block 6: Alternative estimation methods and their implementation.
Hypothesis testing Week 10 Lecture 2.
Chapter Seventeen HYPOTHESIS TESTING
1/55 EF 507 QUANTITATIVE METHODS FOR ECONOMICS AND FINANCE FALL 2008 Chapter 10 Hypothesis Testing.
Topic 2: Statistical Concepts and Market Returns
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 8-1 Business Statistics: A Decision-Making Approach 6 th Edition Chapter.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Basic Business Statistics.
Inference about a Mean Part II
Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Statistics for Business and Economics 7 th Edition Chapter 9 Hypothesis Testing: Single.
Inferences About Process Quality
BCOR 1020 Business Statistics Lecture 18 – March 20, 2008.
Chapter 8 Introduction to Hypothesis Testing
Choosing Statistical Procedures
AM Recitation 2/10/11.
Statistics 11 Hypothesis Testing Discover the relationships that exist between events/things Accomplished by: Asking questions Getting answers In accord.
Aaker, Kumar, Day Ninth Edition Instructor’s Presentation Slides
Probability Distributions and Test of Hypothesis Ka-Lok Ng Dept. of Bioinformatics Asia University.
Overview of Statistical Hypothesis Testing: The z-Test
Chapter 10 Hypothesis Testing
Overview Definition Hypothesis
Confidence Intervals and Hypothesis Testing - II
Business Statistics, A First Course (4e) © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Business Statistics,
Statistical Techniques I
Fundamentals of Hypothesis Testing: One-Sample Tests
1/2555 สมศักดิ์ ศิวดำรงพงศ์
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap th Lesson Introduction to Hypothesis Testing.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
The paired sample experiment The paired t test. Frequently one is interested in comparing the effects of two treatments (drugs, etc…) on a response variable.
1 Power and Sample Size in Testing One Mean. 2 Type I & Type II Error Type I Error: reject the null hypothesis when it is true. The probability of a Type.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Inferential Statistics.
Population All members of a set which have a given characteristic. Population Data Data associated with a certain population. Population Parameter A measure.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 22 Using Inferential Statistics to Test Hypotheses.
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Hypothesis Testing: One Sample Cases. Outline: – The logic of hypothesis testing – The Five-Step Model – Hypothesis testing for single sample means (z.
Chapter 10 Hypothesis Testing
1 Introduction to Hypothesis Testing. 2 What is a Hypothesis? A hypothesis is a claim A hypothesis is a claim (assumption) about a population parameter:
Lecture 7 Introduction to Hypothesis Testing. Lecture Goals After completing this lecture, you should be able to: Formulate null and alternative hypotheses.
1 MAXIMUM LIKELIHOOD ESTIMATION Recall general discussion on Estimation, definition of Likelihood function for a vector of parameters  and set of values.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Biostatistics Class 6 Hypothesis Testing: One-Sample Inference 2/29/2000.
Hypothesis Testing A procedure for determining which of two (or more) mutually exclusive statements is more likely true We classify hypothesis tests in.
Inference and Inferential Statistics Methods of Educational Research EDU 660.
4 Hypothesis & Testing. CHAPTER OUTLINE 4-1 STATISTICAL INFERENCE 4-2 POINT ESTIMATION 4-3 HYPOTHESIS TESTING Statistical Hypotheses Testing.
CHAPTER 9 Testing a Claim
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 8-1 Chapter 8 Fundamentals of Hypothesis Testing: One-Sample Tests Statistics.
Chap 8-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 8 Introduction to Hypothesis.
Interval Estimation and Hypothesis Testing Prepared by Vera Tabakova, East Carolina University.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
Lecture 3: Statistics Review I Date: 9/3/02  Distributions  Likelihood  Hypothesis tests.
Chap 8-1 Fundamentals of Hypothesis Testing: One-Sample Tests.
Ex St 801 Statistical Methods Inference about a Single Population Mean.
Inen 460 Lecture 2. Estimation (ch. 6,7) and Hypothesis Testing (ch.8) Two Important Aspects of Statistical Inference Point Estimation – Estimate an unknown.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
© Copyright McGraw-Hill 2004
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University.
BIOL 582 Lecture Set 2 Inferential Statistics, Hypotheses, and Resampling.
Hypothesis Tests u Structure of hypothesis tests 1. choose the appropriate test »based on: data characteristics, study objectives »parametric or nonparametric.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 7 Inferences Concerning Means.
McGraw-Hill/Irwin © 2003 The McGraw-Hill Companies, Inc.,All Rights Reserved. Part Four ANALYSIS AND PRESENTATION OF DATA.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Chapter 5 STATISTICAL INFERENCE: ESTIMATION AND HYPOTHESES TESTING
Lecture Nine - Twelve Tests of Significance.
Chapter 9 Hypothesis Testing.
STATISTICAL INFERENCE PART IV
Presentation transcript:

DATA ANALYSIS Module Code: CA660 Lecture Block 5

2 Estimator validity – how good? Basis statistical properties (variance, bias, distributional etc.) Bias where is the point estimate,  the true parameter. Bias can be positive, negative or zero. Permits calculation of other properties, e.g. where this quantity and variance of estimator are the same if estimator is unbiased. Obtained by both analytical and “bootstrap methods” Similarly, for continuous variables or for b bootstrap replications, ESTIMATION – Rationale, Summary & Other Features

3 Estimation Rationale etc. - contd. For any, estimator, even unbiased, there is a difference between estimator and true parameter = sampling error Hence the need for probability statements around with C.L. for estimator = (T 1, T 2 ), similarly to before and  the confidence coefficient. If the estimator is unbiased, in other words,  is the probability that the true parameter falls into the interval. In general, confidence intervals can be determined using parametric and non-parametric approaches, where parametric construction needs a pivotal quantity = variable which is a function of parameter and data, but whose distribution does not depend on the parameter.

4 HYPOTHESIS TESTING - Rationale Starting Point of scientific research e.g. No Genetic Linkage between the genetic markers and genes when we design a linkage mapping experiment H 0 :  = 0.5 (No Linkage) (2-locus linkage experiment) H 1 :   0.5 (two loci linked with specified R.F. = 0.2, say) Critical Region Given a cumulative probability distribution fn. of a test statistic, F(x) say, the critical region for the hypothesis test is the region of rejection in the distribution, i.e. the area under the probability curve where the observed test statistic value is unlikely to be observed if H 0 true.  = significance level

5 HT: Critical Regions and Symmetry For a symmetric 2-tailed hypothesis test: or distinction = uni- or bi-directional alternative hypotheses Non-Symmetric, 2-tailed For a=0 or b=0, reduces to 1-tailed case

6 HT-Critical Values and Significance Cut-off values for Rejection and Acceptance regions = Critical Values, so hypothesis test can be interpreted as comparison between critical values and observed hypothesis test statistic, i.e. Significance Level : p-value is the probability of observing a sample outcome if H 0 true is cum. prob. that expected value less than observed test statistic for data under H 0. For p-value less than or equal to , H 0 rejected at significance level  and below.

7 Related issues in Hypothesis Testing - POWER of the TEST Probability of False Positive and False Negative errors e.g. false positive if linkage between two genes declared, when really independent Hypothesis Test Result Fact Accept H 0 Reject H 0 H 0 True 1-  False positive = Type I error =  H 0 False False negative Power of the Test =Type II error=  = 1-  Power of the Test or Statistical Power = probability of rejecting H 0 when correct to do so. (Related strictly to alternative hypothesis and  )

8 Example on Type II Error and Power Suppose have a variable, with known population S.D. = 3.6. From the population, a r.s. size n=100, used to test at  =0.05, that critical values of for a 2-sided test are: for  =0.05 where for, i = upper or lower and  0  under H 0 So substituting our values gives: But, if H 0 false,  is not 17.5, but some other value e.g say ??

9 Example contd. Want new distribution with mean  = 16.5, i.e. new distribution is shifted w.r.t. the old. Thus the probability of the Type II error - failing to reject false H 0 is the area under the curve in the new distribution which overlaps the non-rejection region specified under H 0 So, this is Thus, probability of taking the appropriate action (rejecting H 0 when this is false) is = Power

10 Shifting the distributions Non-Rejection region  /2 Rejection region 

11 Example contd. - Power under alternative  for given  Possible values of   1-  under H 1 for H 0 false Balancing  and  :  tends to be large c.f.  unless original hypothesis way off. So decision based on a rejected H 0 more conclusive than one based on H 0 not rejected, as probability wrong is larger in the latter case.

12 SAMPLE SIZE DETERMINATION Example: Suppose wanted to design a genetic mapping experiment, usually mating design (or population mapping type): so conventional experimental design - ANOVA), genetic marker type and sample size considered. Questions might include: What is the statistical power to detect linkage for certain progeny size? What is the precision of estimated R.F. when sample size is N? Sample size needed for specific Statistical Power Sample size needed for specific Confidence Interval

13 Sample size - calculation based on C.I. For some parameter for normal approximation approach valid, C.I. is U =standardized normal deviate (S.N.D.) and range from lower to upper limits is just a precision measurement for the estimator Given a true parameter , So manipulation gives:

14 Sample size - calculation based on Power (firstly, what affects power)? Suppose  = 0.05,  =3.5, n=100, testing H 0 :  0 =25 when true  =24; assume H 1 :  1 <25 - One-tailed test (U = 1.645):sample mean under H 0 = 24.45: Under H 1  Power = = 0.89 Note: Two-sided test at  = 0.05 gives critical values with under H 0 : equivalently  U L =+0.89,U u = 4.82 for H 1 (i.e.substitute for limits in equn. & then recalculate for new  =  1 = 24) So, P{do not reject H 0 :  =25 when true mean  =24} = =  (Type II) Thus, Power = =

15 Sample Size and Power contd. Suppose, n=25, other values same. 1-tailed now Power = Suppose  = 0.01, critical values 2-tailed with, equivalently, U L = +0.29, U U = So, P{do not reject H 0 :  =25 when true mean  =24} = Power = FACTORS : , n and type of test (1- or 2-sided), true parameter value where subscripts 0 and 1 refer to null and alternative, and  value may refer to 1-sided value or 2-sided value

16 ‘Other’ ways to estimate/test NON-PARAMETRICS/DIST N FREE Standard Pdfs -do not apply to data, sampling distributions or test statistics – uncertain due to small or unreliable data sets, non-independence etc. Parameter estimation - not key issue. Example / Empirical-basis. Weaker assumptions. Less ‘information’ e.g. median used. Simple hypothesis testing as opposed to estimation. Power and efficiency are issues. Counts - nominal, ordinal (natural non-parametric data type). Nonparametric Hypothesis Tests - (parallels to the parametric case). e.g. H.T. of locus orders requires complex test statistic distributions, so need to construct empirical pdf. Usually, assume the null hypothesis using re-sampling techniques, e.g. permutation tests, bootstrap, jacknife.

17 LIKELIHOOD - DEFINITIONS Suppose X can take a set of values x 1,x 2,…with where is a vector of parameters affecting observed x’s e.g.. So can say something about P{X} if we know, say, But not usually case, i.e. observe x’s, knowing nothing of Assuming x’s a random sample size n from a known distribution, then likelihood for Finding most likely for given data is equivalent to Maximising the Likelihood function, (where M.L.E. is )

18 LIKELIHOOD –SCORE and INFO. CONTENT The Log-likelihood is a support function [S(  )] evaluated at point,  ´ say Support function for any other point, say  ´´ can also be obtained – basis for computational iterations for MLE e.g. Newton-Raphson SCORE = first derivative of support function w.r.t. the parameter or, numerically/discretely, INFORMATION CONTENT evaluated at (i) arbitrary point = Observed Info. (ii)support function maximum = Expected Info.

19 Example - Binomial variable (e.g. use of Score, Expected Info. Content to determine type of mapping population and sample size for genomics experiments) Likelihood function Log-likelihood Assume n constant, so first term invariant w.r.t. p = S(  ) at point p Maximising w.r.t. p i.e. set the derivative of S w.r.t. = 0 so SCORE so M.L.E. How does it work, why bother?

20 Numerical Examples: See some data sets and test examples: Simple: od.PDF Context: All sections useful, but especially examples, sections 1-3 and 6 Also, e.g. for R

21 Bayesian Estimation- in context Parametric Estimation - in “classical approach” f(x,  ) for a r.v. X of density f(x), with  the unknown parameter  dependency of distribution on parameter to be estimated. Bayesian Estimation-  is a random variable, so can consider the density as conditional and write f(x|  ) Given a random sample X 1, X 2,… X n the sample random variables are jointly distributed with parameter r.v. . So, joint pdf Objective - to form an estimator that gives value of , dependent on observations of the sample random variables. Thus conditional density of  given X 1, X 2,… X n also plays a role. This is the posterior density

22 Bayes - contd. Posterior Density Relationship - prior and posterior: where  (  ) prior density of  Value: Close to MLE for large n, or for small n if sample values compatible with the prior distribution. Also, has strong sample basis, simpler to calculate than M.L.E.

23 Estimator Comparison in brief. Classical: uses objective probabilities, intuitive estimators, additional assumptions for sampling distributions: good properties for some estimators. Moment:less calculation, less efficient. Not common in genomic analysis despite analytical solutions & low bias, because poorer asymptotic properties; even simple solutions may not be unique. Bayesian - subjective prior knowledge, sample info. close to MLE under certain conditions - see earlier. LSE - if assumptions met,  ’s unbiased + variances obtained, {(X T X) -1 }. Few assumptions for response variable distributions, just expectations, variance-covariance structure. (Unlike MLE where need to specify joint prob. distribution of variables). Requires additional assumptions for sampling distns. Close if assumptions met. Computation easier.

24 Addendum: MGF – estimation use Mostly for functions of Random variables. Normal : reproductive property ( and Poisson and  2 ) actually has Importantly, for X 1, X 2,…X n independent r.v.’s with normal distributions with means and variances Then r.v. is ~ N with