Metrics, Bayes, and BOGSAT: Recognizing and Assessing Uncertainties in Earthquake Hazard Maps Seth Stein 1, Edward M. Brooks 1, Bruce D. Spencer 2 1 Department.

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

TOPIC 3: HOW WELL CAN WE PREDICT EARTHQUAKE HAZARDS? Predictions are important for hazard mitigation policy How much should we believe them?
Polynomial Regression and Transformations STA 671 Summer 2008.
Ground Motions Geotechnical Earthquake Engineering: Steve Kramer
AP STATISTICS Simulation “Statistics means never having to say you're certain.”
Statistical Issues in Research Planning and Evaluation
Sensitivity Analysis for Observational Comparative Effectiveness Research Prepared for: Agency for Healthcare Research and Quality (AHRQ)
Details for Today: DATE:3 rd February 2005 BY:Mark Cresswell FOLLOWED BY:Assignment 2 briefing Evaluation of Model Performance 69EG3137 – Impacts & Models.
International Insurance-Reinsurance Forum 2012 Managing catastrophic risks From Seismic Hazard Analysis to Seismic Risk mitigation C.O.Cioflan Ph.D., Senior.
Since New Madrid's not moving... A complex system view of midcontinental seismicity and hazards Seth Stein Northwestern Eric Calais Purdue Qingsong Li.
Simple Linear Regression
Statistical Methods Chichang Jou Tamkang University.
4. TESTS 5. CONCLUSIONS 2. METHOD We base the study on the method proposed by Zúñiga and Wyss (1995) which was originally intended for finding suitable.
Lecture 10 Comparison and Evaluation of Alternative System Designs.
Time-dependent seismic hazard maps for the New Madrid seismic zone and Charleston, South Carolina areas James Hebden Seth Stein Department of Earth and.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Aaker, Kumar, Day Seventh Edition Instructor’s Presentation Slides
Yan Y. Kagan Dept. Earth and Space Sciences, UCLA, Los Angeles, CA , Global.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
Seth Stein, Earth & Planetary Sciences, Northwestern University
In SAS March 11, Mariya Cheryomina.  power (π) + β = 1  β = (1- π) = probability of accepting false Ho (ie. reject true Ha) - probability of.
Monté Carlo Simulation MGS 3100 – Chapter 9. Simulation Defined A computer-based model used to run experiments on a real system.  Typically done on a.
The Empirical Model Karen Felzer USGS Pasadena. A low modern/historical seismicity rate has long been recognized in the San Francisco Bay Area Stein 1999.
2007 NSTA: St. Louis, Missouri Earthquake Prediction and Forecasting: A Case Study of the San Andreas and New Madrid Faults Sponsored by: IRIS (Incorporated.
Aaker, Kumar, Day Ninth Edition Instructor’s Presentation Slides
Copyright © Cengage Learning. All rights reserved. 8 Tests of Hypotheses Based on a Single Sample.
Bad assumptions or bad luck: Tohoku’s embarrassing lessons for earthquake hazard mapping What’s going wrong and what to do? Tohoku, Japan March 11, 2011.
Introduction In probability, events are either dependent or independent. Two events are independent if the occurrence or non-occurrence of one event has.
2. Shallow versus deep uncertainties Why many predictions / forecasts fail.
VI. Evaluate Model Fit Basic questions that modelers must address are: How well does the model fit the data? Do changes to a model, such as reparameterization,
Research opportunities using IRIS and other seismic data resources John Taber, Incorporated Research Institutions for Seismology Michael Wysession, Washington.
Inductive Generalizations Induction is the basis for our commonsense beliefs about the world. In the most general sense, inductive reasoning, is that in.
Are we successfully addressing the PSHA debate? Seth Stein Earth & Planetary Sciences, Northwestern University.
Earthquakes & Society –tsunami –seismic gap Objectives Discuss factors that affect the amount of damage done by an earthquake. Explain some of the factors.
Accuracy Assessment Having produced a map with classification is only 50% of the work, we need to quantify how good the map is. This step is called the.
Earthquake forecasting using earthquake catalogs.
Earthquake hazard isn’t a physical thing we measure. It's something mapmakers define and then use computer programs to predict. To decide how much to believe.
Managerial Economics Demand Estimation & Forecasting.
Statistical Hypotheses & Hypothesis Testing. Statistical Hypotheses There are two types of statistical hypotheses. Null Hypothesis The null hypothesis,
Time series Model assessment. Tourist arrivals to NZ Period is quarterly.
Hypothesis Testing. The 2 nd type of formal statistical inference Our goal is to assess the evidence provided by data from a sample about some claim concerning.
INTRODUCTION TO Machine Learning 3rd Edition
Yan Y. Kagan Dept. Earth and Space Sciences, UCLA, Los Angeles, CA , Evaluation.
Confidence Interval & Unbiased Estimator Review and Foreword.
9. As hazardous as California? USGS/FEMA: Buildings should be built to same standards How can we evaluate this argument? Frankel et al., 1996.
PCB 3043L - General Ecology Data Analysis.
What to do given that earthquake hazard maps often fail
Department of Mathematics and Geosciences 1 Department of Earth & Planetary Sciences and Institute for Policy Research, Northwestern University, Evanston,
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
1 Ivan Wong Principal Seismologist/Vice President Seismic Hazards Group, URS Corporation Oakland, CA Uncertainties in Characterizing the Cascadia Subduction.
Machine Learning 5. Parametric Methods.
How good can hazard maps be & how good do they need to be Seth Stein, Earth & Planetary Sciences, Northwestern University Jerome Stein, Applied Mathematics,
PART 2 SPSS (the Statistical Package for the Social Sciences)
Migrating earthquakes and faults switching on and off: a new view of intracontinental earthquakes Seth Stein Northwestern University Mian Liu University.
Lessons from Tohoku: why earthquake hazard maps often fail and what to do about it Tohoku, Japan March 11, 2011 M 9.1 NY Times CNN Seth Stein, Northwestern.
Chapter 13 Understanding research results: statistical inference.
9. As hazardous as California? USGS/FEMA: Buildings should be built to same standards How can we evaluate this argument? Frankel et al., 1996.
The Chi Square Equation Statistics in Biology. Background The chi square (χ 2 ) test is a statistical test to compare observed results with theoretical.
1 Department of Earth & Planetary Sciences and Institute for Policy Research, Northwestern University, Evanston, Illinois, USA 2 Department of Statistics.
ChE 551 Lecture 04 Statistical Tests Of Rate Equations 1.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Antonella Peresan F. Vaccari, A. Magrin, G.F. Panza, S. Cozzini B.K. Rastogi, I. Parvez Antonella Peresan F. Vaccari, A. Magrin, G.F. Panza, S. Cozzini.
Outline Historical note about Bayes’ rule Bayesian updating for probability density functions –Salary offer estimate Coin trials example Reading material:
CSC321: Lecture 8: The Bayesian way to fit models Geoffrey Hinton.
Test of a Population Median. The Population Median (  ) The population median ( , P 50 ) is defined for population T as the value for which the following.
Earth & Planetary Sciences, Northwestern University
Why aren't earthquake hazard maps better. Seth Stein1, M
Tohoku earthquake: A surprise?
Discrete Event Simulation - 4
Presentation transcript:

Metrics, Bayes, and BOGSAT: Recognizing and Assessing Uncertainties in Earthquake Hazard Maps Seth Stein 1, Edward M. Brooks 1, Bruce D. Spencer 2 1 Department of Earth & Planetary Sciences and Institute for Policy Research, Northwestern University 2 Department of Statistics and Institute for Policy Research, Northwestern University

Hazard maps often do poorly 2010 map predicts probability of strong shaking in next 30 years But: 2011 M 9.1 Tohoku, 1995 Kobe M 7.3 & others in areas mapped as low hazard Geller 2011

Due to space-time variability of earthquakes & short record, hazard map bull’s-eyes reflect known past earthquakes, may not indicate future ones China GSHAP (1999) map PGA m/s 2 10% in 50 yr (1/500 yr) 2008 Wenchuan

Hazard maps are like ‘Whack-a-mole’ - you wait for the mole to come up where it went down, but it’s likely to pop up somewhere else.

NY Times 3/21/11

Issues Maps often do poorly Large uncertainties unquantified & not communicated to users Revising map after large earthquake may or may not predict future better Map performance not assessed: no agreed criteria for doing so How maps perform is unknown and involves complicated and not understood factors

Hazard maps are hard to get right: successfully predicting future shaking depends on accuracy of four assumptions over years Where will large earthquakes occur? When will they occur? How large will they be? How strong will their shaking be? Try to assess uncertainty via either -sensitivity analysis of effects of poorly-known parameters -evaluation of map performance, hard because of long times required to test forecasts

Newman et al., 2001 PREDICTED HAZARD DEPENDS GREATLY ON - Assumed maximum magnitude of largest events -Assumed ground motion model -Neither are well known since large earthquakes rare 180% 275% 2% in 50 yr (1/2500 yr)

154% %106 PREDICTED HAZARD ALSO DEPENDS ON RECURRENCE MODEL For New Madrid, usual time- independent model predicts higher hazard

Stein et al, 2012 Seismic hazard uncertainty at least factor of 3-4 At best partially reducible because some key parameters poorly known, unknown, or unknowable Large uncertainties not communicated to users

Options after an earthquake yields shaking larger than anticipated: Either regard the high shaking as a low- probability event allowed by the map Or – as usually done - accept that high shaking was not simply a low-probability event and revise the map

No formal or objective criteria used to decide whether to change map & how Done via BOGSAT (“Bunch Of Guys Sitting Around Table”) Challenge: a new map that better describes the past may or may not better predict the future

Deciding whether to remake a map is like deciding after a coin has come up heads a number of times whether to continue assuming that the coin is fair and the run is a low- probability event, or to change to a model in which the coin is assumed to be biased. Changing the model to match past may describe future worse ?

Bayes’ Rule – how much to change depends on one’s confidence in prior model Revised probability model = Likelihood of observations given the prior model x Prior probability model If you were confident that the coin was fair, you would probably not change your model. If you were given the coin at a magic show, your confidence would be lower and you would be more likely to change your model. ?

How good a baseball player was Babe Ruth? The answer depends on the metric used. In some seasons Ruth led the league in both home runs and in the number of times he struck out. By one metric he did very well, and by another, very poorly. Assessing performance

Metrics compare map and observations M0 – Exceedance Metric –M0 = | f – p | –Compares fraction (f) of sites exceeding map to predicted fraction of exceedances (p) - Implicit in PSHA M1 – Squared Misfit Metric –M1 = 1 / N ∑(s i – x i ) 2 –Quantifies magnitude of difference between observation and prediction –similar to visual comparison Predicted p% Observed f % 100-p% 100-f % Each metric gives different insight into map behavior Stein et al., 2015

A map can be nominally successful by the fractional exceedance metric, but significantly underpredict shaking at many sites and overpredict that at others. Or be nominally unsuccessful by the fractional site exceedance metric, but better predict shaking at most sites. p=0.1 f=0.1 M0=0 p=0.1 f=0.2 M0=0.1 underpredicted overpredicted

The short time period since hazard maps began to be made is a challenge for assessing how well they work. Hindcasts offer long records, but are not true tests, as they compare maps to data that were available when the map was made. Still, they give useful insight. Compare 510-year shaking record to Japanese National Hazard (JNH) maps Brooks et al., 2015 Miyazawa and Mori, 2009

Uniform & random maps Geller (2011) argued that “all of Japan is at risk from earthquakes, and the present state of seismological science does not allow us to reliably differentiate the risk level in particular geographic areas,” so a map showing uniform hazard would be preferable. Test: By the exceedance metric (M0) uniform and randomized maps do better, but by the square misfit metric (M1) JNH maps perform better. Brooks et al., 2015 JNH worse JNH better

Smoothed Maps Observed Original 50 km 100 km How detailed should a map be? –By exceedance metric (M0) map improves by smoothing over larger areas –By squared misfit (M1), map performs best if smoothed over km Brooks et al., 2015

Maps may be overparameterized (overfit). A high order polynomial fits past data better than linear or quadratic models, but this more detailed model predicts the future worse than the simpler models. Intermediate level of detail may be better for hazard maps.

Conclusions Maps often do poorly Large uncertainties unquantified & not communicated to users Map performance not assessed: no agreed criteria for doing so In some cases, by some metrics, uniform, random, or smoothed maps perform better Map performance involves complicated and not yet understood factors Better understanding of performance & uncertainties could improve hazard policy use