X-ray Astrostatistics Bayesian Methods in Data Analysis Aneta Siemiginowska Vinay Kashyap and CHASC Jeremy Drake, Nov.2005.

Slides:



Advertisements
Similar presentations
Bayesian Hypothesis Testing In Nested Models Harold Jeffreys Jeff Rouder.
Advertisements

Anthony Greene1 Simple Hypothesis Testing Detecting Statistical Differences In The Simplest Case:  and  are both known I The Logic of Hypothesis Testing:
Inspiral Parameter Estimation via Markov Chain Monte Carlo (MCMC) Methods Nelson Christensen Carleton College LIGO-G Z.
Deriving and fitting LogN-LogS distributions An Introduction Andreas Zezas University of Crete.
Chap 9: Testing Hypotheses & Assessing Goodness of Fit Section 9.1: INTRODUCTION In section 8.2, we fitted a Poisson dist’n to counts. This chapter will.
Markov-Chain Monte Carlo
AP Statistics – Chapter 9 Test Review
X-Ray Spectroscopy Workshop Cambridge, MA Matthew Carpenter, UCB SSL 7/11/2007, Comparison of Observed and Theoretical Fe L Emission from CIE Plasmas Matthew.
Bayesian inference Gil McVean, Department of Statistics Monday 17 th November 2008.
Hypothesis testing Week 10 Lecture 2.
Computing the Posterior Probability The posterior probability distribution contains the complete information concerning the parameters, but need often.
GLAST LAT Project Astrostatistics Workshop, HEAD meeting, 10 September 2004 James Chiang (GSSC/UMBC) 1 Gamma-ray Large Area Space Telescope Challenges.
Constraining Astronomical Populations with Truncated Data Sets Brandon C. Kelly (CfA, Hubble Fellow, 6/11/2015Brandon C. Kelly,
Chapter Seventeen HYPOTHESIS TESTING
Section 7.1 Hypothesis Testing: Hypothesis: Null Hypothesis (H 0 ): Alternative Hypothesis (H 1 ): a statistical analysis used to decide which of two competing.
Basic Elements of Testing Hypothesis Dr. M. H. Rahbar Professor of Biostatistics Department of Epidemiology Director, Data Coordinating Center College.
Hypothesis : Statement about a parameter Hypothesis testing : decision making procedure about the hypothesis Null hypothesis : the main hypothesis H 0.
Course overview Tuesday lecture –Those not presenting turn in short review of a paper using the method being discussed Thursday computer lab –Turn in short.
Deriving and fitting LogN-LogS distributions Andreas Zezas Harvard-Smithsonian Center for Astrophysics.
G. Cowan RHUL Physics Bayesian Higgs combination page 1 Bayesian Higgs combination using shapes ATLAS Statistics Meeting CERN, 19 December, 2007 Glen Cowan.
Statistical hypothesis testing – Inferential statistics I.
AM Recitation 2/10/11.
Bayesian Model Comparison and Occam’s Razor Lecture 2.
Overview of Statistical Hypothesis Testing: The z-Test
Chapter 8 Introduction to Hypothesis Testing
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
Calibration Guidelines 1. Start simple, add complexity carefully 2. Use a broad range of information 3. Be well-posed & be comprehensive 4. Include diverse.
Three data analysis problems Andreas Zezas University of Crete CfA.
Model Inference and Averaging
Chapter 9 Hypothesis Testing II: two samples Test of significance for sample means (large samples) The difference between “statistical significance” and.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Significance Toolbox 1) Identify the population of interest (What is the topic of discussion?) and parameter (mean, standard deviation, probability) you.
Inference and Inferential Statistics Methods of Educational Research EDU 660.
A taste of statistics Normal error (Gaussian) distribution  most important in statistical analysis of data, describes the distribution of random observations.
A statistical test for point source searches - Aart Heijboer - AWG - Cern june 2002 A statistical test for point source searches Aart Heijboer contents:
On Predictive Modeling for Claim Severity Paper in Spring 2005 CAS Forum Glenn Meyers ISO Innovative Analytics Predictive Modeling Seminar September 19,
Power and Sample Size Anquan Zhang presents For Measurement and Statistics Club.
Cosmological Model Selection David Parkinson (with Andrew Liddle & Pia Mukherjee)
Ch15: Decision Theory & Bayesian Inference 15.1: INTRO: We are back to some theoretical statistics: 1.Decision Theory –Make decisions in the presence of.
How to handle calibration uncertainties in high-energy astrophysics Vinay Kashyap SPIE 2008 :: Observatory Operations :: :: 16:20 Wednesday 25.
The White Dwarf Age of NGC 2477 Elizabeth Jeffery Space Telescope Science Institute Collaborators: Ted von Hippel, Steven DeGennaro, David van Dyk, Nathan.
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
© Copyright McGraw-Hill 2004
Metal abundance evolution in distant galaxy clusters observed by XMM-Newton Alessandro Baldi Astronomy Dept. - University of Bologna INAF - OABO In collaboration.
Inference ConceptsSlide #1 1-sample Z-test H o :  =  o (where  o = specific value) Statistic: Test Statistic: Assume: –  is known – n is “large” (so.
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
G. Cowan Computing and Statistical Data Analysis / Stat 9 1 Computing and Statistical Data Analysis Stat 9: Parameter Estimation, Limits London Postgraduate.
2005 Unbinned Point Source Analysis Update Jim Braun IceCube Fall 2006 Collaboration Meeting.
G. Cowan, RHUL Physics Statistics for early physics page 1 Statistics jump-start for early physics ATLAS Statistics Forum EVO/Phone, 4 May, 2010 Glen Cowan.
ELEC 303 – Random Signals Lecture 17 – Hypothesis testing 2 Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 2, 2009.
1 Probability and Statistics Confidence Intervals.
Statistical Methods. 2 Concepts and Notations Sample unit – the basic landscape unit at which we wish to establish the presence/absence of the species.
Chapter 13 Understanding research results: statistical inference.
In Bayesian theory, a test statistics can be defined by taking the ratio of the Bayes factors for the two hypotheses: The ratio measures the probability.
Hypothesis Testing. Statistical Inference – dealing with parameter and model uncertainty  Confidence Intervals (credible intervals)  Hypothesis Tests.
Hypothesis Tests u Structure of hypothesis tests 1. choose the appropriate test »based on: data characteristics, study objectives »parametric or nonparametric.
1 Interpretation of the Test Statistic or: basic Hypothesis Testing, with applications, in 15 minutes Patrick Nolan Stanford University GLAST LAT DC2 Kickoff.
McGraw-Hill/Irwin © 2003 The McGraw-Hill Companies, Inc.,All Rights Reserved. Part Four ANALYSIS AND PRESENTATION OF DATA.
Chapter 9: Hypothesis Tests for One Population Mean 9.5 P-Values.
Chapter Nine Hypothesis Testing.
Lecture 1.31 Criteria for optimal reception of radio signals.
Dr.MUSTAQUE AHMED MBBS,MD(COMMUNITY MEDICINE), FELLOWSHIP IN HIV/AIDS
AstroStatistics Workshop
Special Topics In Scientific Computing
اختبار الفرضيات اختبارالفرضيات المتعلقة بالوسط
Statistical Inference: One- Sample Confidence Interval
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Appetizer: Statistical methods in XSPEC
CS639: Data Management for Data Science
Inference as Decision Section 10.4.
Presentation transcript:

X-ray Astrostatistics Bayesian Methods in Data Analysis Aneta Siemiginowska Vinay Kashyap and CHASC Jeremy Drake, Nov.2005

X-ray Astrostatistics Bayesian Methods in Data Analysis Aneta Siemiginowska Vinay Kashyap and CHASC Jeremy Drake, Nov.2005

CHASC: California-Harvard Astrostatistics Collaboration History: why this collaboration? Regular Seminars: each second Tuesday at the Science Center Participate in SAMSI workshop => Spring 2006 Participants: HU Statistics Dept., Irvine UC, and CfA astronomers Topics related mostly to X-ray astronomy, but also sun-spots! Papers: MCMC for X-ray data, Fe-line and F-test issues, EMC2, hardness ratio and line detection Algorithms are described in the papers => working towards public release Stat: David van Dyk, Xiao-Li Meng, Taeyoung Park, Yaming Yu, Rima Izem Astro: Alanna Connors, Peter Freeman, Vinay Kashyap, Aneta Siemiginowska Andreas Zezas, James Chiang, Jeff Scargle

X-ray Data Analysis and Statistics Different type analysis: Spectral, image, timing. XSPEC and Sherpa provide the main fitting/modeling environments X-ray data => counting photons: -> normal - Gaussian distribution for high number of counts, but very often we deal with low counts data Low counts data (< 10) => Poisson data and  2 is not appropriate! Several modifications to  2 have been developed: Weighted  2 (.e.g. Gehrels 1996) Formulation of Poisson Likelihood (  C follows  for N>5) Cash statistics: (Cash 1979) C-statistics - goodness-of-fit and background (in XSPEC, Keith Arnaud)

Steps in Data Analysis Obtain data - observations! Reduce - processing the data, extract image, spectrum etc. Analysis - Fit the data Conclude - Decide on Model, Hypothesis Testing! Reflect

Hypothesis Testing How to decide which model is better? A simple power law or blackbody? A simple power law or continuum with emission lines? Statistically decide: how to reject a simple model and accept more complex one? Standard (Frequentist!) Model Comparison Tests: Goodness-of-fit Maximum Likelihood Ratio test F-test

Steps in Hypothesis Testing - I

Steps in Hypothesis Testing - II Two model Mo (simpler) and M1 (more complex) were fit to the data D; Mo => null hypothesis. Construct test statistics T from the best fit of two models: e.g.   =       Determine each sampling distribution for T statistics, e.g. p(T | Mo) and p(T | M1) Determine significance  => Reject Mo when p (T | Mo) <  Determine the power of the test =>  probability of selecting Mo when M1 is correct p(T|Mo) p(T|M1)

Conditions for LRT and F-test The two models that are being compared have to be nested: broken power law is an example of a nested model BUT power law and thermal plasma models are NOT nested The null values of the additional parameters may not be on the boundary of the set of possible parameter values: continuum + emission line -> line intensity = 0 on the boundary References Freeman et al 1999, ApJ, 524, 753 Protassov et al 2002, ApJ 571, 545

Simple Steps in Calibrating the Test: 1.Simulate N data sets (e.g. use fakeit in Sherpa or XSPEC): => the null model with the best-fit parameters (e.g. power law, thermal) => the same background, instrument responses, exposure time as in the initial analysis 2.(A) Fit the null and alternative models to each of the N simulated data sets and (B) compute the test statistic: T LRT = -2log [L(   |sim)/L(   |sim)]       best fit parameters  T F =     3.Compute the p-value - proportion of simulations that results in a value of statistic (T) more extreme than the value computed with the observed data. p-value = (1/N) * Number of [ T(sim) > T(data) ]

Simulation Example M0 - power law M1 - pl+narrow line M2 - pl+broad line M3 - pl+absorption line M0/M1 M0/M2M0/M3 Comparison between p-value And significance in the    distribution  =0.05    Reject Null Accept Null

Simulation Example M0 - power law M1 - pl+narrow line M2 - pl+broad line M3 - pl+absorption line M0/M1 M0/M2M0/M3 Comparison between p-value And significance in the    distribution  =0.05    Reject Null Accept Null

Bayesian Methods use Bayesian approach - max likelihood, priors, posterior distribution - to fit/find the modes of the posterior (best fit parameters) Simulate from the posterior distribution, including uncertainties on the best-fit parameters, Calculate posterior predictive p-values Bayes factors: direct comparison of probabilities P(M1)/P(Mo)

CHASC Projects at SAMSI 2006 Source and Feature detection Working group Issues in Modeling High Counts Data Image reconstructions (e.g. Solar data) Detection and upper limits in high background data (GLAST) Smoothed/unsharp mask images - significance of features Issues in Low Counts Data Upper limits Classification of Sources - point source vs. extended Poisson data in the presence of Poisson Background Quantification of uncertainty and Confidence Other Projects in Town: Calibration uncertainties in X-ray analysis Emission Measure model for X-ray spectroscopy (Log N - Log S) model in X-ray surveys

Bayesian Model Comparison To compare two models, a Bayesian computes the odds, or odd ratio: where B 10 is the Bayes factor. When there is no a priori preference for either model, B 10 = 1 of one indicates that each model is equally likely to be correct, while B 10 ≥ 10 may be considered sufficient to accept the alternative model (although that number should be greater if the alternative model is controversial).

Model Comparison Tests A model comparison test statistic T is created from the best-fit statistics of each fit; it is sampled from a probability distribution p(T). The test significance is defined as the integral of p(T) from the observed value of T to infinity. The significance quantifies the probability that one would select the more complex model when in fact the null hypothesis is correct. A standard threshold for selecting the more complex model is significance < 0.05 (the "95% criterion" of statistics). p(T|Mo) p(T|M1)