Maximum Likelihood See Davison Ch. 4 for background and a more thorough discussion. Sometimes.

Slides:



Advertisements
Similar presentations
STA305 Spring 2014 This started with excerpts from STA2101f13
Advertisements

Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
Introduction to Regression with Measurement Error STA431: Spring 2015.
Exploratory Factor Analysis
The Art of Model Building and Statistical Tests. 2 Outline The art of model building Using Software output The t-statistic The likelihood ratio test The.
1 1 Slide © 2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
Likelihood Ratio, Wald, and Lagrange Multiplier (Score) Tests
Statistical Estimation and Sampling Distributions
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
Hypothesis: It is an assumption of population parameter ( mean, proportion, variance) There are two types of hypothesis : 1) Simple hypothesis :A statistical.
Chapter 13: The Chi-Square Test
Regression Part II One-factor ANOVA Another dummy variable coding scheme Contrasts Multiple comparisons Interactions.
Logistic Regression STA302 F 2014 See last slide for copyright information 1.
Maximum likelihood (ML) and likelihood ratio (LR) test
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 14 Goodness-of-Fit Tests and Categorical Data Analysis.
Chi Square Distribution (c2) and Least Squares Fitting
Maximum likelihood (ML)
Introduction to Regression with Measurement Error STA431: Spring 2013.
The Chi-square Statistic. Goodness of fit 0 This test is used to decide whether there is any difference between the observed (experimental) value and.
AP STATISTICS LESSON 13 – 1 (DAY 1) CHI-SQUARE PROCEDURES TEST FOR GOODNESS OF FIT.
Introduction to Regression with Measurement Error STA302: Fall/Winter 2013.
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Multinomial Distribution
Logistic Regression STA2101/442 F 2014 See last slide for copyright information.
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
Copyright © 2009 Cengage Learning 15.1 Chapter 16 Chi-Squared Tests.
Chapter 26 Chi-Square Testing
Regression Part II One-factor ANOVA Another dummy variable coding scheme Contrasts Multiple comparisons Interactions.
Introduction Many experiments result in measurements that are qualitative or categorical rather than quantitative. Humans classified by ethnic origin Hair.
Maximum Likelihood Estimation Methods of Economic Investigation Lecture 17.
Chi- square test x 2. Chi Square test Symbolized by Greek x 2 pronounced “Ki square” A Test of STATISTICAL SIGNIFICANCE for TABLE data.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
Fitting probability models to frequency data. Review - proportions Data: discrete nominal variable with two states (“success” and “failure”) You can do.
Confirmatory Factor Analysis Part Two STA431: Spring 2013.
Categorical Independent Variables STA302 Fall 2013.
Copyright (C) 2002 Houghton Mifflin Company. All rights reserved. 1 Understandable Statistics S eventh Edition By Brase and Brase Prepared by: Lynn Smith.
STA302: Regression Analysis. Statistics Objective: To draw reasonable conclusions from noisy numerical data Entry point: Study relationships between variables.
Copyright © 2010 Pearson Education, Inc. Warm Up- Good Morning! If all the values of a data set are the same, all of the following must equal zero except.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Chapter 5 Statistical Inference Estimation and Testing Hypotheses.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 11 Analyzing the Association Between Categorical Variables Section 11.2 Testing Categorical.
Chapter 15 The Chi-Square Statistic: Tests for Goodness of Fit and Independence PowerPoint Lecture Slides Essentials of Statistics for the Behavioral.
Statistics 300: Elementary Statistics Section 11-2.
M.Sc. in Economics Econometrics Module I Topic 4: Maximum Likelihood Estimation Carol Newman.
Chapter 13- Inference For Tables: Chi-square Procedures Section Test for goodness of fit Section Inference for Two-Way tables Presented By:
Logistic Regression Saed Sayad 1www.ismartsoft.com.
Categorical Independent Variables STA302 Fall 2015.
Chapter 14 – 1 Chi-Square Chi-Square as a Statistical Test Statistical Independence Hypothesis Testing with Chi-Square The Assumptions Stating the Research.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 12 Tests of Goodness of Fit and Independence n Goodness of Fit Test: A Multinomial.
STA302: Regression Analysis. Statistics Objective: To draw reasonable conclusions from noisy numerical data Entry point: Study relationships between variables.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University.
The Sample Variation Method A way to select sample size This slide show is a free open source document. See the last slide for copyright information.
Logistic Regression For a binary response variable: 1=Yes, 0=No This slide show is a free open source document. See the last slide for copyright information.
Richard Kass/F02P416 Lecture 6 1 Lecture 6 Chi Square Distribution (  2 ) and Least Squares Fitting Chi Square Distribution (  2 ) (See Taylor Ch 8,
Chapter 11: Categorical Data n Chi-square goodness of fit test allows us to examine a single distribution of a categorical variable in a population. n.
CHI SQUARE DISTRIBUTION. The Chi-Square (  2 ) Distribution The chi-square distribution is the probability distribution of the sum of several independent,
Log-linear Models Please read Chapter Two. We are interested in relationships between variables White VictimBlack Victim White Prisoner151 (151/160=0.94)
Chapter 4. Inference about Process Quality
Math 4030 – 10b Inferences Concerning Variances: Hypothesis Testing
Poisson Regression STA 2101/442 Fall 2017
Test for Goodness of Fit
Interactions and Factorial ANOVA
Active Learning Lecture Slides
Logistic Regression Part One
See Davison Ch. 4 for background and a more thorough discussion.
Likelihood Ratio, Wald, and Lagrange Multiplier (Score) Tests
Econ 3790: Business and Economics Statistics
POINT ESTIMATOR OF PARAMETERS
Exploratory Factor Analysis. Factor Analysis: The Measurement Model D1D1 D8D8 D7D7 D6D6 D5D5 D4D4 D3D3 D2D2 F1F1 F2F2.
Presentation transcript:

Maximum Likelihood See Davison Ch. 4 for background and a more thorough discussion. Sometimes

Maximum Likelihood Sometimes

Close your eyes and differentiate?

Simulate Some Data: True α=2, β=3 Alternatives for getting the data into D might be D = scan(“Gamma.data”) -- Can put entire URL D = c(20.87, 13.74, …, 10.94)

Log Likelihood

R function for the minus log likelihood

Where should the numerical search start? How about Method of Moments estimates? E(X) = αβ, Var(X) = αβ 2 Replace population moments by sample moments and put a ~ above the parameters.

If the second derivatives are continuous, H is symmetric. If the gradient is zero at a point and |H|≠0, If H is positive definite, local minimum If H is negative definite, local maximum If H has both positive and negative eigenvalues, saddle point

A slicker way to define the minus log likelihood function

Likelihood Ratio Tests Under H 0, G 2 has an approximate chi-square distribution for large N. Degrees of freedom = number of (non-redundant, linear) equalities specified by H 0. Reject when G 2 is large.

Example: Multinomial with 3 categories Parameter space is 2-dimensional Unrestricted MLE is (P 1, P 2 ): Sample proportions. H 0 : θ 1 = 2θ 2

Parameter space and restricted parameter space

R code for the record

Degrees of Freedom Express H 0 as a set of linear combinations of the parameters, set equal to constants (usually zeros). Degrees of freedom = number of non-redundant linear combinations (meaning linearly independent). df=3 (count the = signs)

Can write Null Hypothesis in Matrix Form as H 0 : Lθ = h

Gamma Example: H 0 : α = β

Make a wrapper function

It's probably okay, but plot -LL

Test H 0 : α = β

The actual Theorem (Wilks, 1934) There are r+p parameters Null hypothesis says that the first r parameters equal specified constants. Then under some regularity conditions, G 2 converges in distribution to chi- squared with r df if H 0 is true. Can justify tests of linear null hypotheses by a re-parameterization using the invariance principle.

How it works The invariance principle of maximum likelihood estimation says that the MLE of a function is that function of the MLE. Like Meaning is particularly clear when the function is one-to-one. Write H 0 : Lθ = h, where L is r x (r+p) and rows of L are linearly independent. Can always find an additional p vectors that, together with the rows of L, span R r+p This defines a (linear) 1-to-1 re-parameterization, and Wilks' theorem applies directly.

Gamma Example H 0 : α = β

Can Work for Non-linear Null Hypotheses Too

Copyright Information This slide show was prepared by Jerry Brunner, Department of Statistics, University of Toronto. It is licensed under a Creative Commons Attribution - ShareAlike 3.0 Unported License. Use any part of it as you like and share the result freely. These Powerpoint slides will be available from the course website: