EC 331 The Theory of and applications of Maximum Likelihood Method

Slides:



Advertisements
Similar presentations
Tests of Static Asset Pricing Models
Advertisements

Generalized Method of Moments: Introduction
Copula Regression By Rahul A. Parsa Drake University &
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
2 – In previous chapters: – We could design an optimal classifier if we knew the prior probabilities P(wi) and the class- conditional probabilities P(x|wi)
Likelihood Ratio, Wald, and Lagrange Multiplier (Score) Tests
SOLVED EXAMPLES.
Chap 8: Estimation of parameters & Fitting of Probability Distributions Section 6.1: INTRODUCTION Unknown parameter(s) values must be estimated before.
Visual Recognition Tutorial
Maximum likelihood (ML) and likelihood ratio (LR) test
Maximum likelihood Conditional distribution and likelihood Maximum likelihood estimations Information in the data and likelihood Observed and Fisher’s.
Maximum likelihood (ML)
AGC DSP AGC DSP Professor A G Constantinides© Estimation Theory We seek to determine from a set of data, a set of parameters such that their values would.
Maximum likelihood (ML) and likelihood ratio (LR) test
Maximum Likelihood We have studied the OLS estimator. It only applies under certain assumptions In particular,  ~ N(0, 2 ) But what if the sampling distribution.
Today Today: Chapter 9 Assignment: 9.2, 9.4, 9.42 (Geo(p)=“geometric distribution”), 9-R9(a,b) Recommended Questions: 9.1, 9.8, 9.20, 9.23, 9.25.
Maximum-Likelihood estimation Consider as usual a random sample x = x 1, …, x n from a distribution with p.d.f. f (x;  ) (and c.d.f. F(x;  ) ) The maximum.
Visual Recognition Tutorial
Continuous Random Variables and Probability Distributions
Pattern Recognition Topic 2: Bayes Rule Expectant mother:
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Maximum likelihood (ML)
Random Sampling, Point Estimation and Maximum Likelihood.
Lecture 3: Inference in Simple Linear Regression BMTRY 701 Biostatistical Methods II.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Maximum Likelihood Estimation Methods of Economic Investigation Lecture 17.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
: Chapter 3: Maximum-Likelihood and Baysian Parameter Estimation 1 Montri Karnjanadecha ac.th/~montri.
Lecture 3: Statistics Review I Date: 9/3/02  Distributions  Likelihood  Hypothesis tests.
1 Standard error Estimated standard error,s,. 2 Example 1 While measuring the thermal conductivity of Armco iron, using a temperature of 100F and a power.
M.Sc. in Economics Econometrics Module I Topic 4: Maximum Likelihood Estimation Carol Newman.
CLASSICAL NORMAL LINEAR REGRESSION MODEL (CNLRM )
Lecture 1: Basic Statistical Tools. A random variable (RV) = outcome (realization) not a set value, but rather drawn from some probability distribution.
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
Review of statistical modeling and probability theory Alan Moses ML4bio.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
MathematicalMarketing Slide 5.1 OLS Chapter 5: Ordinary Least Square Regression We will be discussing  The Linear Regression Model  Estimation of the.
The Probit Model Alexander Spermann University of Freiburg SS 2008.
Conditional Expectation
Presentation : “ Maximum Likelihood Estimation” Presented By : Jesu Kiran Spurgen Date :
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
MathematicalMarketing Slide 3c.1 Mathematical Tools Chapter 3: Part c – Parameter Estimation We will be discussing  Nonlinear Parameter Estimation  Maximum.
The Probit Model Alexander Spermann University of Freiburg SoSe 2009
Large Sample Distribution Theory
The Maximum Likelihood Method
Large Sample Theory EC 532 Burak Saltoğlu.
STATISTICS POINT ESTIMATION
12. Principles of Parameter Estimation
Parameter Estimation 主講人:虞台文.
Inference for the mean vector
The Maximum Likelihood Method
The Maximum Likelihood Method
Large Sample Theory EC 532 Burak Saltoğlu.
More about Posterior Distributions
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
The Multivariate Normal Distribution, Part 2
Integration of sensory modalities
Computing and Statistical Data Analysis / Stat 7
Parametric Methods Berlin Chen, 2005 References:
Learning From Observed Data
STATISTICAL INFERENCE PART I POINT ESTIMATION
12. Principles of Parameter Estimation
Maximum Likelihood We have studied the OLS estimator. It only applies under certain assumptions In particular,  ~ N(0, 2 ) But what if the sampling distribution.
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
HKN ECE 313 Exam 2 Review Session
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Applied Statistics and Probability for Engineers
Maximum Likelihood Estimation (MLE)
Presentation transcript:

EC 331 The Theory of and applications of Maximum Likelihood Method Burak Saltoğlu

outline Maximum Liklelihood Principle Estimating population parameters via ML method Properties of ML OLS vs ML

This represents the joint density of y’s given parameter 1 Maximum Likelihood ML method is based on the principle that the parameter estimates can be obtained by maximising the likelihood of the selected sample to reflect the population. We choose the parameters in a way that we maximize the joint likelihood of representing the population. Suppose we are given iid observed sample of y and also a parameter vector (of k dimesion) can be represented as This represents the joint density of y’s given parameter

Likelihood Function Joint likelihood function then can be written as the joint probability of observing y’s drawn from f(.) Likelihood function is Maximizing above function w.r.t. will yield a special value that maximizes the probability of obtaining sample values that have actually observed. In most applications it is convinient to work with loglikelihood function, which is

Likelihood Function Note that Also note that above equation is known as score .

Example-1 Poisson distribbution due to Siméon Denis Poisson expresses the probability of a given number of events occurring in a fixed interval of time these events occur with a known average rate and independently of the time since the last event use: defaults of countries, customers,

Example-1

Example-1

Numerical example

Numerical example

Likelihood profile (lambda in the horizontal axis)

Likelihood and log-likelihood for Poisson (rescaled Graph)

Example-2 It describes the time between events in a Poisson process

Example-2

Example-2

Example-3

Example-4

Convergence in Probability Definition : Let xn be a sequence random variable where n is sample size, the random variable xn converges in probability to a constant c if the values that the x may take that are not close to c become increasingly unlikely as n increases. If xn converges to c, then we say, All the mass of the probability distribution concentrates around c.

Properties of MLE Consistency: Asymtotic Normality: where information matrix is that is, the hessian of log-likelihood function.

3.3 Properties of MLE Asymtotic Efficiency: Assumimg that we are dealing with only one parameter θ; which states that if there is another consistent and asymtotically normal estimator of to θ then , Invariance:

4 Estimation of the Linear Regression Model

Matrix notation

3.4 Estimation of the Linear Regression Model Parameter vector is

3.4 Estimation of the Linear Regression Model To calculate variance matrix of parameters, we need hessian of likelihood parameters. İf we take ot second derivatives Taking expectations,

3.4 Estimation of the Linear Regression Model

3.4 Estimation of the Linear Regression Model So, the information matrix is The inverse of the information matrix will give us the variance-covariance matrix of the MLE estimators,

Testing in Maximum Likelihood Framework

Example from Poisson example

Example This ratio is always between 0 and 1 and the less likely the assumption is, the smaller this ratio 

Likelihood Ratio Test If we want to test Restricted likelihood ratio defined as can be used with decision rule Restricted Unrestricted q:#restrictions

Likelihood Ratio Test Don’t reject the null

More on LR test in the context of Linear Regression

Likelihood Ratio Test