Comments on Hierarchical models, and the need for Bayes Peter Green, University of Bristol, UK IWSM, Chania, July 2002.

Slides:



Advertisements
Similar presentations
Introduction to Monte Carlo Markov chain (MCMC) methods
Advertisements

Previous Lecture: Distributions. Introduction to Biostatistics and Bioinformatics Estimation I This Lecture By Judy Zhong Assistant Professor Division.
1 Some Comments on Sebastiani et al Nature Genetics 37(4)2005.
Mean, Proportion, CLT Bootstrap
Week 11 Review: Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
Psychology 290 Special Topics Study Course: Advanced Meta-analysis April 7, 2014.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Flipping A Biased Coin Suppose you have a coin with an unknown bias, θ ≡ P(head). You flip the coin multiple times and observe the outcome. From observations,
Sampling Distributions (§ )
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
Chapter 7 Sampling and Sampling Distributions
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 9: Hypothesis Tests for Means: One Sample.
1 BA 555 Practical Business Analysis Review of Statistics Confidence Interval Estimation Hypothesis Testing Linear Regression Analysis Introduction Case.
STA Lecture 201 STA 291 Lecture 20 Exam II Today 5-7pm Memorial Hall (Same place as exam I) Makeup Exam 7:15pm – 9:15pm Location CB 234 Bring a calculator,
Standard error of estimate & Confidence interval.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 12 Analyzing the Association Between Quantitative Variables: Regression Analysis Section.
Review of Basic Statistics. Definitions Population - The set of all items of interest in a statistical problem e.g. - Houses in Sacramento Parameter -
Additional Slides on Bayesian Statistics for STA 101 Prof. Jerry Reiter Fall 2008.
Sampling Theory Determining the distribution of Sample statistics.
The paired sample experiment The paired t test. Frequently one is interested in comparing the effects of two treatments (drugs, etc…) on a response variable.
F OUNDATIONS OF S TATISTICAL I NFERENCE. D EFINITIONS Statistical inference is the process of reaching conclusions about characteristics of an entire.
STA Lecture 161 STA 291 Lecture 16 Normal distributions: ( mean and SD ) use table or web page. The sampling distribution of and are both (approximately)
Bayesian Inference, Basics Professor Wei Zhu 1. Bayes Theorem Bayesian statistics named after Thomas Bayes ( ) -- an English statistician, philosopher.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 8-1 Confidence Interval Estimation.
Random Sampling, Point Estimation and Maximum Likelihood.
PARAMETRIC STATISTICAL INFERENCE
Stat 13, Tue 5/8/ Collect HW Central limit theorem. 3. CLT for 0-1 events. 4. Examples. 5.  versus  /√n. 6. Assumptions. Read ch. 5 and 6.
Part 0 -- Introduction Statistical Inference and Regression Analysis: Stat-GB , C Professor William Greene Stern School of Business IOMS.
Bayes for Beginners Reverend Thomas Bayes ( ) Velia Cardin Marta Garrido.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
1 Chapter 10: Introduction to Inference. 2 Inference Inference is the statistical process by which we use information collected from a sample to infer.
Bayesian Classification. Bayesian Classification: Why? A statistical classifier: performs probabilistic prediction, i.e., predicts class membership probabilities.
Confidence Intervals: The Basics BPS chapter 14 © 2006 W.H. Freeman and Company.
Stat 112: Notes 2 Today’s class: Section 3.3. –Full description of simple linear regression model. –Checking the assumptions of the simple linear regression.
Introduction to Inferece BPS chapter 14 © 2010 W.H. Freeman and Company.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Sampling distributions rule of thumb…. Some important points about sample distributions… If we obtain a sample that meets the rules of thumb, then…
Confidence Interval & Unbiased Estimator Review and Foreword.
: An alternative representation of level of significance. - normal distribution applies. - α level of significance (e.g. 5% in two tails) determines the.
Ch15: Decision Theory & Bayesian Inference 15.1: INTRO: We are back to some theoretical statistics: 1.Decision Theory –Make decisions in the presence of.
Inference: Probabilities and Distributions Feb , 2012.
Inen 460 Lecture 2. Estimation (ch. 6,7) and Hypothesis Testing (ch.8) Two Important Aspects of Statistical Inference Point Estimation – Estimate an unknown.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Math 3680 Lecture #15 Confidence Intervals. Review: Suppose that E(X) =  and SD(X) = . Recall the following two facts about the average of n observations.
Regression Analysis1. 2 INTRODUCTION TO EMPIRICAL MODELS LEAST SQUARES ESTIMATION OF THE PARAMETERS PROPERTIES OF THE LEAST SQUARES ESTIMATORS AND ESTIMATION.
1 Probability and Statistics Confidence Intervals.
CSE 473 Uncertainty. © UW CSE AI Faculty 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one has stood.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Sampling Theory Determining the distribution of Sample statistics.
The accuracy of averages We learned how to make inference from the sample to the population: Counting the percentages. Here we begin to learn how to make.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Chapter 10 Confidence Intervals for Proportions © 2010 Pearson Education 1.
Chapter Nine Hypothesis Testing.
Statistical Inference
Inference for the Mean of a Population
Inference: Conclusion with Confidence
Bayes Net Learning: Bayesian Approaches
Inference for Proportions
CI for μ When σ is Unknown
Bayesian Inference, Basics
Introduction to Probability & Statistics The Central Limit Theorem
Statistical NLP: Lecture 4
Essential Statistics Two-Sample Problems - Two-sample t procedures -
Reasoning in Psychology Using Statistics
CS 594: Empirical Methods in HCC Introduction to Bayesian Analysis
Sampling Distributions (§ )
CS639: Data Management for Data Science
How Confident Are You?.
Presentation transcript:

Comments on Hierarchical models, and the need for Bayes Peter Green, University of Bristol, UK IWSM, Chania, July 2002

Complex data structures Multiple sources of variability >1 strata Measurement error, indirect observation Random effects, latent variables Hierarchical population structure (multi- level models) Experimental regimes, missing data

Complex data structures, ctd. …. all features prevalent in complex biomedical data, especially  need for Hierarchical Models e.g. many talks here at IWSM17 generalised linear models are just not enough (and that’s not because of linearity or exponential families)

Inference in hierarchical models it is important to plug-in estimates generally lead to under- estimating variability of quantities of interest - AVOID! we need a coherent calculus of uncertainty propagate all sources of variability

a coherent calculus of uncertainty? we have one - it’s called Probability!  full probability modelling of all variables reported inference: joint distribution of unknowns of interest, given observed data how? Bayes’ theorem Inference in hierarchical models, ctd.

Costs and benefits costs –more modelling work –computational issues (?) benefits – valid analysis –avoiding ad-hoc decisions –counts all data once and once only!

Costs and benefits, ctd. by-product – simultaneous, coherent inference about multiple targets and the old question: what about sensitivity to prior assumptions? –if sensitivity analysis reveals strong dependence on prior among reasonable prior choices, how can you trust the non- Bayesian analysis?

A simple prediction problem (an example that plugging in is wrong) We make 10 observations; their mean is 15 and standard deviation 2. What is the chance that the next observation will be more than 19?

… prediction, continued Can’t do much without assumptions - let’s suppose the data are normal…….. …. 19 is 2 s.d.’s more than the mean, and the normal distribution probability of that is 2.3%

… prediction, continued But this supposes that 15 and 2 are the population mean and s.d. We ought to allow for our uncertainty in these numbers - they are only estimates This is awkward to do for a non- Bayesian

… prediction, continued The Bayesian answer - (1) if the mean  and s.d.  were known, the answer would be 1-  ((19-  )/  ) (2) we should average this quantity over the posterior distribution of ( ,  ) - I did this and got 4.5% - twice the ‘plug-in’ answer!

Summary (1) Bayes inference is completely sound mathematically - ‘coherent’ All your conclusions are self-consistent Handles prediction properly Allows sequential updating No logical somersaults (confidence intervals, hypothesis tests) Bayes estimators are often more accurate

Summary (2) But it does require more input than just the data Sensitivity to priors should be checked Computation is an issue except in very simple problems - that’s true for non- Bayes too

Reading Migon, H.S. and Gamerman, D. Statistical Inference: an integrated approach. Arnold, Box, G.E.P. and Tiao, G.C. Bayesian inference in Statistical Analysis. Addison-Wesley, Carlin, B.P. and Louis, T.A. Bayes and empirical Bayes methods for data analysis. Chapman and Hall, Gelman, A., Carlin, J.B., Stern, H.S. and Rubin, D.B. Bayesian data analysis. Chapman and Hall, 1995.

Professor Peter Green Department of Mathematics University of Bristol Bristol BS8 1TW, UK tel: fax: 7999