Chapter 5 Statistical Inference Estimation and Testing Hypotheses.

Slides:



Advertisements
Similar presentations
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 9 Inferences Based on Two Samples.
Advertisements

General Linear Model With correlated error terms  =  2 V ≠  2 I.
A. The Basic Principle We consider the multivariate extension of multiple linear regression – modeling the relationship between m responses Y 1,…,Y m and.
Point Estimation Notes of STAT 6205 by Dr. Fan.
Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
Lecture XXIII.  In general there are two kinds of hypotheses: one concerns the form of the probability distribution (i.e. is the random variable normally.
Ch8 Inference concerning variance
Chapter 7. Statistical Estimation and Sampling Distributions
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
Chap 9: Testing Hypotheses & Assessing Goodness of Fit Section 9.1: INTRODUCTION In section 8.2, we fitted a Poisson dist’n to counts. This chapter will.
Chap 8: Estimation of parameters & Fitting of Probability Distributions Section 6.1: INTRODUCTION Unknown parameter(s) values must be estimated before.
The General Linear Model. The Simple Linear Model Linear Regression.
Hypothesis testing Some general concepts: Null hypothesisH 0 A statement we “wish” to refute Alternative hypotesisH 1 The whole or part of the complement.
Minimaxity & Admissibility Presenting: Slava Chernoi Lehman and Casella, chapter 5 sections 1-2,7.
Probability theory 2011 The multivariate normal distribution  Characterizing properties of the univariate normal distribution  Different definitions.
Chapter Goals After completing this chapter, you should be able to:
Inferences About Means of Single Samples Chapter 10 Homework: 1-6.
2. Point and interval estimation Introduction Properties of estimators Finite sample size Asymptotic properties Construction methods Method of moments.
Chapter 11 Multiple Regression.
Chapter 3 Hypothesis Testing. Curriculum Object Specified the problem based the form of hypothesis Student can arrange for hypothesis step Analyze a problem.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 14 Goodness-of-Fit Tests and Categorical Data Analysis.
Chapter 2 Simple Comparative Experiments
Lecture 7 1 Statistics Statistics: 1. Model 2. Estimation 3. Hypothesis test.
Inferences About Process Quality
Probability theory 2008 Outline of lecture 5 The multivariate normal distribution  Characterizing properties of the univariate normal distribution  Different.
The Multivariate Normal Distribution, Part 2 BMTRY 726 1/14/2014.
Chapter 7 Inferences Regarding Population Variances.
Multivariate Analysis of Variance, Part 1 BMTRY 726.
Maximum Likelihood Estimation
Hypothesis Testing and T-Tests. Hypothesis Tests Related to Differences Copyright © 2009 Pearson Education, Inc. Chapter Tests of Differences One.
METU Informatics Institute Min 720 Pattern Classification with Bio-Medical Applications PART 2: Statistical Pattern Classification: Optimal Classification.
Aaker, Kumar, Day Ninth Edition Instructor’s Presentation Slides
Inference for the mean vector. Univariate Inference Let x 1, x 2, …, x n denote a sample of n from the normal distribution with mean  and variance 
Principles of Pattern Recognition
Statistical Decision Theory
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 11 Inferences About Population Variances n Inference about a Population Variance n.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
Chapter 7 Point Estimation
Lecture 3: Statistics Review I Date: 9/3/02  Distributions  Likelihood  Hypothesis tests.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Brief Review Probability and Statistics. Probability distributions Continuous distributions.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Brief Review Probability and Statistics. Probability distributions Continuous distributions.
© Copyright McGraw-Hill 2004
Charles University FSV UK STAKAN III Institute of Economic Studies Faculty of Social Sciences Institute of Economic Studies Faculty of Social Sciences.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Stats & Summary. The Woodbury Theorem where the inverses.
n Point Estimation n Confidence Intervals for Means n Confidence Intervals for Differences of Means n Tests of Statistical Hypotheses n Additional Comments.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University.
Week 21 Order Statistics The order statistics of a set of random variables X 1, X 2,…, X n are the same random variables arranged in increasing order.
Chapter 8 Estimation ©. Estimator and Estimate estimator estimate An estimator of a population parameter is a random variable that depends on the sample.
Chapter 11: Categorical Data n Chi-square goodness of fit test allows us to examine a single distribution of a categorical variable in a population. n.
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 7 Inferences Concerning Means.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Probability Theory and Parameter Estimation I
Math 4030 – 10b Inferences Concerning Variances: Hypothesis Testing
Inferences Regarding Population Variances
Elementary Statistics
Statistical Inference about Regression
The Multivariate Normal Distribution, Part 2
Addition of Independent Normal Random Variables
EM for Inference in MV Data
Inferences Regarding Population Variances
Parametric Methods Berlin Chen, 2005 References:
EM for Inference in MV Data
Chapter 8 Estimation.
Statistical Inference for the Mean: t-test
Fundamental Sampling Distributions and Data Descriptions
Presentation transcript:

Chapter 5 Statistical Inference Estimation and Testing Hypotheses

5.1 Data Sets & Matrix Normal Distribution Data matrix where n rows X 1, …, X n are iid

Vec(X') is an np×1 random vector with We writeMore general, we can define matrix normal distribution.

Definition An n×p random matrix X is said to follow a matrix normal distributionif where In this case, where W=BB', V=AA', Y has i.i.d. elements each following N(0,1).

Theorem The density function of with W > 0, V >0 is given by where etr(A)= exp(tr(A)). Corollary 1: Let X be a matrix of n observations from Then the density function of X is where

5.2 Maximum Likelihood Estimation A. Review Step 1. The likelihood function

Step 2. Domain (parameter space) The MLE ofmaximizes over H.

Step 3. Maximization

Results 4.9 (p168 of textbook )

B. Multivariate population Step 1. The likelihood function Step 2. Domain

Step 3. Maximization (a) We can prove that P(B > 0) = 1 if n > p.

(b) We have

(c) Let λ 1, …, λ p be the eigenvalues of Σ *. The function g(λ)= λ -n/2 e -1/ 2λ arrives its maximum at λ=1/n. The function L(Σ * ) arrives its maximum at λ 1 =1/n, …, λ p =1/n and (d) The MLE of Σ is

Theorem Let X 1, …, X n be a sample from with n > p and. Then the MLEs of are respectively, and the maximum likelihood is

Theorem Under the above notations, we have a) are independent; b) c) is a biased estimator of A unbiased estimator of is recommended by called the sample covariance matrix.

Matalb code: mean, cov, corrcoef Theorem Let be the MLE of and be a measurable function. Then is the MLE of. Corollary 1 The MLE of the correlations is

5.3 Wishart distribution A. Chi-square distribution Let X 1, …, X n are iid N(0,1). Then, the chi-square distribution with n degrees of freedom or Definition If x ~ N n (0, I n ), then Y= x ' x is said to have a chi-square distribution with n degrees of freedom, and write.

B. Wishart distribution (obtained by Wishart in 1928) Definition Let. Then we said that W= x ' x is distributed according to a Wishart distribution.

A.Unbiaseness Let be an estimator of. If is called unbiased estimator of. 5.4 Discussion on estimation Theorem Let X 1, …, X n be a sample from Then are unbiased estimators of and, respectively. Matlab code: mean, cov, corrcoef

B. Decision Theory Then the average of loss is give by That is called the risk function.

Definition An estimator t(X) is called a minimax estimator of if Example 1 Under the loss function the sample mean is a minimax estimator of.

C. Admissible estimation Definition An estimator t 1 (x) is said to be at least as good as another t 2 (x) if And t 1 is said to be better than or strictly dominates t 2 if the above inequality holds with strict inequality for at least one.

Definition An estimator t * is said to be inadmissible if there exists another estimator t ** that is better than t *. An estimator t * is admissible if it is not inadmissible.  The admissibility is a weak requirement.  Under the loss, the sample mean is an inadmissible if the population is  James & Stein pointed out is better than The estimator is called James-Stein estimator.

5.5 Inferences about a mean vector (Ch.5 Textbook) Let X 1, …, X n be iid samples from Case A: is known. a)p = 1 b)p > 1

Under the hypothesis H 0, Then Theorem Let X 1, …, X n be a sample from where is known. The null distribution of under is and the rejection area is

Case B: is unknown. a)Suggestion: Replaceby the Sample Covariance Matrix S in, i.e. where Likelihood Ratio Criterion. There are many theoretic approaches to find a suitable statistic. One of the methods is the Likelihood Ratio Criterion.

The Likelihood Ratio Criterion (LRC) Step 1The likelihood function Step 2Domains

Step 3Maximization We have obtained By a similar way we can find where under

Then, the LRC is Note

Finally Remark: Let t(x) be a statistic for the hypothesis and f(u) is a strictly monotone function. Then is a statistic which is equivalent to t(x). We write

5.6 T 2 -statistic Definition Letandbe independent with n > p. The distribution of is called T 2 distribution. The distribution T 2 is independent of, we shall write As

And Theorem Theorem The distribution of is invariant under all affine transformations of the observations and the hypothesis

Confidence Region A 100 (1- )% confidence region for the mean of a p- dimensional normal distribution is the ellipsoid determined by all such that

Proof: X 1, …, X n

Example (Example 5.2 in Textbook) Perspiration from 20 healthy females was analysis.

Computer calculations provide: and

We evaluate Comparing the observedwith the critical value we see thatand consequently, we reject H 0 at the 10% level of significance.

Definition Let x and y be samples of a population G with mean and covariance matrix The quadratic forms are called Mahalanobis distance (M-distance) between x and y, and x and G, respectively. Mahalanobis Distance

If can be verified that

5.7 Two Samples Problems (Section 6.3, Textbook)

We have two samples from the two populations where are unknown. The LRC is where

Under the hypothesis The confidence region of is where

Example 5.7.1(p ) Jolicoeur and Mosimann (1960) studied the relationship of size and shape for painted turtles. The following table contains their measurements on the carapaces of 24 female and 24 male turtles.

5.8 Multivariate Analysis of Variance A.Review There are k normal populations One wants to test equality of the means

The analysis of variance employs decomposition of sum squares where The testing statistics is

B.Multivariate population (pp ) is unknown, one wants to test

I. The likelihood ratio criterion Step 1The likelihood function Step 2The domains

Step 3Maximization where are the total sum of squares and products matrix and the error sum of squares and products matrix, respectively.

The treatment sum of squares and product matrix The LRC

Definition Assumeare independent, where. The distribution is called Wilks -distribution and write. Theorem Under H 0 we have 1) 2) 3) E and B are independent

Special cases of the Wilks -distributions See pp , Textbook for example.

2.Union-Intersection Decision Rule Consider projection hypothesis

For projection data, we have and the F-statistic With the rejection region The rejection region for H 0 is that implies the testing statistic is or

Lemma 1Let A be a symmetric matrix of order p. Denote by, the eigenvalues of A, and, the associated eigenvectors of A. Then

Lemma 2Let A and B are two p× p matrices and A’ = A, B>0. Denote by and, the eigenvalues of and associated eigenvectors. Then Remark1: Remark2: Remark3: Let be eigenvalues of. The Wilks -statistic can be expressed as