458 Model Uncertainty and Model Selection Fish 458, Lecture 13.

Slides:



Advertisements
Similar presentations
Probability models- the Normal especially.
Advertisements

Tests of Hypotheses Based on a Single Sample
CHAPTER 21 Inferential Statistical Analysis. Understanding probability The idea of probability is central to inferential statistics. It means the chance.
From the homework: Distribution of DNA fragments generated by Micrococcal nuclease digestion mean(nucs) = bp median(nucs) = 110 bp sd(nucs+ = 17.3.
Review You run a t-test and get a result of t = 0.5. What is your conclusion? Reject the null hypothesis because t is bigger than expected by chance Reject.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 9 Hypothesis Testing Developing Null and Alternative Hypotheses Developing Null and.
Day 6 Model Selection and Multimodel Inference
Chap 9: Testing Hypotheses & Assessing Goodness of Fit Section 9.1: INTRODUCTION In section 8.2, we fitted a Poisson dist’n to counts. This chapter will.
Model assessment and cross-validation - overview
Bayesian inference Gil McVean, Department of Statistics Monday 17 th November 2008.
Copyright © 2008 by the McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Managerial Economics, 9e Managerial Economics Thomas Maurice.
458 More on Model Building and Selection (Observation and process error; simulation testing and diagnostics) Fish 458, Lecture 15.
Hypothesis Testing Steps of a Statistical Significance Test. 1. Assumptions Type of data, form of population, method of sampling, sample size.
458 Fitting models to data – II (The Basics of Maximum Likelihood Estimation) Fish 458, Lecture 9.
458 Population Projections (policy analysis) Fish 458; Lecture 21.
Chapter Sampling Distributions and Hypothesis Testing.
BCOR 1020 Business Statistics Lecture 21 – April 8, 2008.
Inference about a Mean Part II
Copyright © 2014, 2013, 2010 and 2007 Pearson Education, Inc. Chapter Hypothesis Tests Regarding a Parameter 10.
Inferences About Process Quality
458 Fish Spring 2005 Fisheries Stock Assessments (Models for Conservation)
BCOR 1020 Business Statistics
Today Concepts underlying inferential statistics
458 Fitting models to data – I (Sum of Squares) Fish 458, Lecture 7.
Correlation and Regression Analysis
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 8 Tests of Hypotheses Based on a Single Sample.
Regression and Correlation Methods Judy Zhong Ph.D.
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Bayesian Model Comparison and Occam’s Razor Lecture 2.
Copyright © Cengage Learning. All rights reserved. 8 Tests of Hypotheses Based on a Single Sample.
The Method of Likelihood Hal Whitehead BIOL4062/5062.
Fundamentals of Data Analysis Lecture 4 Testing of statistical hypotheses.
Topics: Statistics & Experimental Design The Human Visual System Color Science Light Sources: Radiometry/Photometry Geometric Optics Tone-transfer Function.
Lecture 4 Model Selection and Multimodel Inference.
West Virginia University Towards Practical Software Reliability Assessment for IV&V Projects B. Cukic, E. Gunel, H. Singh, V. Cortellessa Department of.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
Chapter 16 The Chi-Square Statistic
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Copyright © 2005 by the McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Managerial Economics Thomas Maurice eighth edition Chapter 4.
Confidence intervals and hypothesis testing Petter Mostad
11 Chapter 12 Quantitative Data Analysis: Hypothesis Testing © 2009 John Wiley & Sons Ltd.
Lecture 16 Section 8.1 Objectives: Testing Statistical Hypotheses − Stating hypotheses statements − Type I and II errors − Conducting a hypothesis test.
Chap 6 Further Inference in the Multiple Regression Model
Chapter 15 The Chi-Square Statistic: Tests for Goodness of Fit and Independence PowerPoint Lecture Slides Essentials of Statistics for the Behavioral.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
G. Cowan Computing and Statistical Data Analysis / Stat 9 1 Computing and Statistical Data Analysis Stat 9: Parameter Estimation, Limits London Postgraduate.
Course: Research in Biomedicine and Health III Seminar 5: Critical assessment of evidence.
Information criteria What function fits best? The more free parameters a model has the higher will be R 2. The more parsimonious a model is the lesser.
Jump to first page Inferring Sample Findings to the Population and Testing for Differences.
Hypothesis Testing. Statistical Inference – dealing with parameter and model uncertainty  Confidence Intervals (credible intervals)  Hypothesis Tests.
Chapter 7: Hypothesis Testing. Learning Objectives Describe the process of hypothesis testing Correctly state hypotheses Distinguish between one-tailed.
Chapter 11: Categorical Data n Chi-square goodness of fit test allows us to examine a single distribution of a categorical variable in a population. n.
Quiz 3. Model selection Overview Objectives determine the “choice” of model Modeling for forecasting Likelihood ratio test Akaike Information Criterion.
A.P. STATISTICS EXAM REVIEW TOPIC #2 Tests of Significance and Confidence Intervals for Means and Proportions Chapters
Model Comparison. Assessing alternative models We don’t ask “Is the model right or wrong?” We ask “Do the data support a model more than a competing model?”
CSC321: Lecture 8: The Bayesian way to fit models Geoffrey Hinton.
Model Comparison.
Chapter 4: Basic Estimation Techniques
Chapter 4 Basic Estimation Techniques
Chapter 24 Comparing Means.
Lecture 4 Model Selection and Multimodel Inference
Basic Estimation Techniques
Review You run a t-test and get a result of t = 0.5. What is your conclusion? Reject the null hypothesis because t is bigger than expected by chance Reject.
Chapter 6 Predicting Future Performance
Basic Estimation Techniques
Hypothesis Tests for a Population Mean in Practice
Model Comparison.
Chapter 6 Predicting Future Performance
1 Chapter 8: Introduction to Hypothesis Testing. 2 Hypothesis Testing The general goal of a hypothesis test is to rule out chance (sampling error) as.
Lecture 4 Model Selection and Multimodel Inference
Presentation transcript:

458 Model Uncertainty and Model Selection Fish 458, Lecture 13

458 Overview Models are hypotheses regarding “how the world could work”. There are usually several competing models ranging from very simple to very complicated. Some important results (e.g. extinction risk – do we have environmental variation in deaths?) will be sensitive to model structure. Complex models explain the data better but may provide poor forecasts. Classical statistics emphasizes estimation uncertainty. However, many would argue that model uncertainty is more important in practice (e.g. which of the data series for northern cod should have been used for assessment purposes).

458 Complexity vs Simplicity-I We wish to approximate a function using a histogram based on 100 points. How many bins should we choose? Too many – imprecise. Too few - biased

458 Complexity vs Simplicity-II Too few parameters, we can’t capture the true model adequately – error due to approximation. Too many parameters, we can’t estimate them adequately – error due to estimation. The “optimal” number of parameters depends on the amount of data.

458 Complexity vs Simplicity-III Consider approximating N(100,25 2 ) using a histogram. We define a “discrepancy” between the predicted and true distributions using:

458 Complexity vs Simplicity-IV The optimal number of bins increases with N

458 Model Selection Model selection can be seen as evaluating the weight of evidence in favor of each hypothesis and using this to select among the hypotheses.

458 Model Selection (Nested Models) A model is nested within another model if it is a special case of that model, e.g. We can compare nested models (model B is nested within model A) using the likelihood ratio test: R, the likelihood ratio, is  2 distributed with number of degrees of freedom equal to the difference in parameters between models A and B.

458 Back to cod-I Some alternative hypotheses : The Base case model (1) is nested within models 2 and 3. Models 4 and 5 are nested within model 1.

458 Back to cod-II Models (A/B) 2 difference  2 -critical (0.05) 2/ / / / Log-likelihood (not the negative-log-likelhood)

458 Model Selection (non-nested models) The likelihood ratio test can only be applied to compare nested models. However, we often wish to compare non-nested models. We use the Akaike Information Criterion (AIC) to make such comparisons. We compute the AIC (AIC c for small sample sizes) for each model and choose that which has the lowest AIC.

458 Model Selection (non-nested models) Choose the model with the lowest value of AIC. Note that the data, Y, are the same for all models.

458 Comparing Growth Curves-I We wish to compare between the von Bertalanffy and logistic growth curves for some simulated data (the true model is the von Bertalanffy curve). We generate 100 data sets based on the von Bertalanffy growth curve for various values for  and count the fraction of cases the von Bertalnffy curve is chosen correctly.

458 Comparing Growth Curves-II Likelihoods (p=4): Von Bert = Logistic = CV=0.2

458 Comparing Growth Curves-III The probability of correctly selecting the von Bertalanffy growth curve depends on  (and the sample size). Checking the reliability of model selection methods by simulation is often worth doing. 

458 Model Selection – Miscellany-I All model selection methods are based on the assumption that the likelihood function is correct. This may well not be the case. Neither likelihood ratio nor AIC can be used to compare models that have different likelihood functions / use different data. Check the residuals about the fits to the data for all models – it may be that none of the models are fitting the data. Model selection makes little sense if none of the models fit the data.

458 Model Selection – Miscellany - II Rejecting models is not always a sensible thing to do. In some cases (e.g. examining the consequences of future management actions), consideration should be given to retaining complicated models even if they don’t provide “significant” improvements in fit. “Model averaging” (e.g. giving a weight to each model – say proportional to exp(AIC)) allows consideration of model uncertainty.

458 Model Selection – Miscellany - III Always plot the fits of the different models. Even if one model is significantly better than another, the improved fit may be qualitatively “insubstantial”. Some models that fit the data better do not provide more “realistic” results (e.g. estimating M often leads to values for M of 0). Likelihood ratio and AIC are frequentist approaches. Bayesian techniques are also available for model selection.

458 Readings Hilborn and Mangel, Chapter 7 Haddon, Chapter 3 Linhart and Zuchini (1986) Burnham and Anderson (1998). Quinn and Deriso, Section 4.5