DATA ANALYSIS: THEORY AND PRACTICE by Vicki M. Bier.

Slides:



Advertisements
Similar presentations
Modelling unknown errors as random variables Thomas Svensson, SP Technical Research Institute of Sweden, a statistician working with Chalmers and FCC in.
Advertisements

1 Component Design Basis Inspection (CDBI) Graydon Strong 6/17/14.
Introduction to Probability and Probabilistic Forecasting L i n k i n g S c i e n c e t o S o c i e t y Simon Mason International Research Institute for.
“ Building Strong “ Delivering Integrated, Sustainable, Water Resources Solutions Expert-Opinion Elicitation Robert C. Patev North Atlantic Division –
Topic 6: Introduction to Hypothesis Testing
EPIDEMIOLOGY AND BIOSTATISTICS DEPT Esimating Population Value with Hypothesis Testing.
UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering On-line Alert Systems for Production Plants A Conflict Based Approach.
© 2005 The McGraw-Hill Companies, Inc., All Rights Reserved. Chapter 5 Making Systematic Observations.
CHAPTER 6 Statistical Analysis of Experimental Data
8-2 Basics of Hypothesis Testing
How Science Works Glossary AS Level. Accuracy An accurate measurement is one which is close to the true value.
Chapter 9 Flashcards. measurement method that uses uniform procedures to collect, score, interpret, and report numerical results; usually has norms and.
Classification of Instruments :
Item Analysis: Classical and Beyond SCROLLA Symposium Measurement Theory and Item Analysis Modified for EPE/EDP 711 by Kelly Bradley on January 8, 2013.
Uncertainties in Thermal Barrier Coating Life Prediction by Karl A. Mentz A Thesis Submitted to the Graduate Faculty of Rensselaer Polytechnic Institute.
TECHNICAL ASSOCIATION OF THE EUROPEAN NATURAL GAS INDUSTRY Development of a Eurogas-Marcogaz Methodology for Estimation of Methane Emissions Angelo Riva.
V. Rouillard  Introduction to measurement and statistical analysis ASSESSING EXPERIMENTAL DATA : ERRORS Remember: no measurement is perfect – errors.
© 2008 McGraw-Hill Higher Education The Statistical Imagination Chapter 9. Hypothesis Testing I: The Six Steps of Statistical Inference.
Copyright © 2006 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide
Development of An ERROR ESTIMATE P M V Subbarao Professor Mechanical Engineering Department A Tolerance to Error Generates New Information….
EVAL 6970: Cost Analysis for Evaluation Dr. Chris L. S. Coryn Nick Saxton Fall 2014.
Epidemiology The Basics Only… Adapted with permission from a class presentation developed by Dr. Charles Lynch – University of Iowa, Iowa City.
VTT-STUK assessment method for safety evaluation of safety-critical computer based systems - application in BE-SECBS project.
ERT 312 SAFETY & LOSS PREVENTION IN BIOPROCESS RISK ASSESSMENT Prepared by: Miss Hairul Nazirah Abdul Halim.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
University of Ottawa - Bio 4118 – Applied Biostatistics © Antoine Morin and Scott Findlay 08/10/ :23 PM 1 Some basic statistical concepts, statistics.
“ Building Strong “ Delivering Integrated, Sustainable, Water Resources Solutions 1 What is Engineering Risk and Reliability? Why We Use It? Robert C.
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
Quality Control Lecture 5
West Virginia University Towards Practical Software Reliability Assessment for IV&V Projects B. Cukic, E. Gunel, H. Singh, V. Cortellessa Department of.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
J1879 Robustness Validation Hand Book A Joint SAE, ZVEI, JSAE, AEC Automotive Electronics Robustness Validation Plan The current qualification and verification.
For ABA Importance of Individual Subjects Enables applied behavior analysts to discover and refine effective interventions for socially significant behaviors.
Therapeutic Equivalence & Active Control Clinical Trials Richard Simon, D.Sc. Chief, Biometric Research Branch National Cancer Institute.
Bayesian vs. frequentist inference frequentist: 1) Deductive hypothesis testing of Popper--ruling out alternative explanations Falsification: can prove.
Lecture 16 Section 8.1 Objectives: Testing Statistical Hypotheses − Stating hypotheses statements − Type I and II errors − Conducting a hypothesis test.
Risk management and disaster preparedness
Statistical Inference for the Mean Objectives: (Chapter 9, DeCoursey) -To understand the terms: Null Hypothesis, Rejection Region, and Type I and II errors.
Academic Research Academic Research Dr Kishor Bhanushali M
Uncertainty Management in Rule-based Expert Systems
Idaho RISE System Reliability and Designing to Reduce Failure ENGR Sept 2005.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Fall 2002Biostat Statistical Inference - Proportions One sample Confidence intervals Hypothesis tests Two Sample Confidence intervals Hypothesis.
Experimental Research Methods in Language Learning Chapter 5 Validity in Experimental Research.
Utility Engineers, PC.  Generation  Transmission  Distribution.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
IMPORTANCE OF STATISTICS MR.CHITHRAVEL.V ASST.PROFESSOR ACN.
IAEA International Atomic Energy Agency INTRODUCTION Module 2.
1 SMU EMIS 7364 NTU TO-570-N Control Charts Basic Concepts and Mathematical Basis Updated: 3/2/04 Statistical Quality Control Dr. Jerrell T. Stracener,
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
Bayesian Approach Jake Blanchard Fall Introduction This is a methodology for combining observed data with expert judgment Treats all parameters.
Statistical NLP: Lecture 4 Mathematical Foundations I: Probability Theory (Ch2)
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Uncertainty and confidence Although the sample mean,, is a unique number for any particular sample, if you pick a different sample you will probably get.
Fundamentals of Data Analysis Lecture 4 Testing of statistical hypotheses pt.1.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Failure Modes, Effects and Criticality Analysis
Chapter Nine Hypothesis Testing.
1 2 3 INDIAN INSTITUTE OF TECHNOLOGY ROORKEE PROJECT REPORT 2016
of Heritage and New Hardware For Launch Vehicle Reliability Models
ACCURACY IN PERCENTILES
Chapter Three Research Design.
GE 6757 TOTAL QUALITY MANAGEMENT
HMI Reliability Dale Wolfe Reliability Engineer LMSSC*ATC*LMSAL
Statistical NLP: Lecture 4
Failure Mode and Effect Analysis
Type I and Type II Errors
Definitions Cumulative time to failure (T): Mean life:
Mikael Olsson Control Engineer
Presentation transcript:

DATA ANALYSIS: THEORY AND PRACTICE by Vicki M. Bier

PRA DATA REQUIREMENTS FOR QUANTIFICATION Initiating event frequencies (internal and external) Component failure rates Component maintenance frequencies and durations Human error rates Common cause failure parameters Component fragilities (thermal, seismic, fire, flood, and missile)

DATA ANALYSIS TOPICS Assembling failure and success data Bayes’ Theorem and Bayesian updating Data analysis for common cause failures

ASSEMBLING FAILURE AND SUCCESS DATA Data collection Data interpretation Categorization of components and failure modes Assembling success (i.e., exposure) data

PRELIMINARY STEPS Determine level of detail –E.g., controller boards versus individual electronic components Identify components of interest –E.g., fuel system isolation valve Determine failure modes –E.g., failure to open on demand, failure to close on demand, plugging during system operation Define data base study period –E.g., all flight data, or only after initial design evolution period

DATA COLLECTION Correlate available data with risk assessment requirements Collect all relevant information –Corrective action reports –Anomaly reports –Operational histories (e.g., mission logs) –Results of check-out, qualification, and acceptance tests

TYPE OF INFORMATION AVAILABLE 1.General engineering knowledge about the design, manufacture, and operation of the equipment in question. 2.The historical performance of the equipment under conditions other than the one under consideration—e.g., test results; generic reliability data from MIL-HDBK; Rome Air Force Base reliability data for non-electronic parts, etc. 3.The historical performance of the equipment under the specific conditions being analyzed.

TYPES OF INFORMATION NEEDED FOR FAILURE FREQUENCY EVALUATION Generic information –Expert opinion –Case histories –Failure frequencies for similar components Component-specific information –Component type –Usage (e.g., cycling, standby, or continuous operation) –Testing, monitoring, and refurbishment policies –Performance history

DATA INTERPRETATION Need to determine: –Appropriate level of detail (e.g., resistors versus circuit boards) –Applicability of test data (e.g., bench tests) –Treatment of incident failures (e.g., turbine blade cracking) and out-of-spec operation (e.g., erratic turbine speed) –Treatment of non-critical components (e.g., temperature or pressure transducers, redundant equipment) –Treatment of corrective actions (e.g., improved cleanliness procedures, system design changes)

APPROPRIATE LEVEL OF DETAIL In general, components should be modeled at the highest level of detail for which good data is available Going beyond this level can significantly decrease accuracy Example – aggregating reliability data for individual electronic components may not match the observed failure history for a controller board

APPLICABILITY OF TEST DATA Tests of installed components under normal operating conditions (e.g., hot firing tests) are likely to give valid results Bench tests or tests outside the normal operating envelope may not give representative results

TREATMENT OF INCIPIENT FAILURES AND OUT-OF-SPECIFICATION OPERATION Examples – turbine blade cracking, erratic turbine speed In general, partial failures should not be included in the data base if the component in question could still perform its intended function Alternative approaches: –Explicitly model the chance of an incipient failure progressing to an actual failure; e.g., turbine blade cracking versus turbine failure –Take partial failures into account in establishing prior distributions; e.g., higher estimated prior probability of turbine over-speed for turbines with a history of erratic speed control

TREATMENT OF NONCRITICAL COMPONENTS Examples – pressure transducers, redundant valves in parallel Non-critical components can still be important: –If another component fails –If some action needs to be taken –Etc. In general, components should be ignored only if they have no safety-related function

TREATMENT OF CORRECTIVE ACTIONS Some types of corrective actions (e.g., addition of a new safety system) may make previous failures inapplicable. However, the effectiveness of actions such as improved maintenance or better cleanliness procedures is difficult to predict without either explicit data or detailed risk models The effectiveness of corrective actions may be disproved by subsequent failures!

CATEGORIZATION OF COMPONENTS AND FAILURE MODES Should similar components be grouped? What is the appropriate level of detail in identifying failure modes?

GROUPING OF SIMILAR COMPONENTS If components are effectively identical, grouping will result in smaller uncertainties by increasing the amount of data available This may not be appropriate due to: –Different operating environments; e.g., identical valves in gas, fuel, and water systems –Different operating modes; e.g., cyclic, continuous, or standby (continuously pulsing valves versus isolation valves) –Different failure models; e.g., fail open versus fail closed (primary versus secondary fuel control valves on loss of power supply) –Different inspection intervals or maintenance policies; e.g., inspected after every flight or not (fuel filters) –Different failure histories

CATEGORIZATION OF FAILURE MODES Treat startup failures differently from run-time failures; e.g., seizing of gearbox Different failure modes must be identified if they have different effects on the system; e.g., valve leaks versus valve stuck in wide open position Different failure causes (e.g., corrosion, fatigue, etc.) do not need to be identified except to aid in prioritization of corrective actions

NEED FOR SUCCESS OR EXPOSURE DATA Failure frequency = Failure of operating systems or components – e.g., loss of flow in a lube oil system (3 events in 110 hours of operation) Failure of passive systems or components – e.g., nitrogen leakage from fuel tank (0 events in 11,000 hours of experience) Failure of standby or cyclic components – e.g., heater switches (0 events in 72 demands) Number of failures Number of tests or operating hours

QUANTIFYING THE RELEVANT EXPOSURE DATA Identify the population of similar components –E.g., how many similar valves or switches are potentially vulnerable to this failure mode? Determine the appropriate units –Does the failure mode occur over time, or only with the imposition of a demand? –Can the failure mode occur at any time, or only when the system is running? –Are there “latent” failures? (e.g., failures occurring during orbit that are only detected at descent)

BAYES’ THEOREM AND BAYESIAN UPDATING

PROPERTIES OF THE BAYESIAN APPROACH 1.All types of information can be used 2.The use of judgment is visible and explicit 3.Provides a way of quantifying uncertainty

BAYESIAN SPECIALIZATION (UPDATE) Update generic distribution of information (“prior distribution”) with more specific or more detailed data (“evidence”) to obtain a weighted distribution which generally contains less uncertainty than the generic information (“specialized posterior distribution”) POSTERIOR PRIOR FAILURE RATE.01%1%

PROPERTIES OF BAYESIAN UPDATING Posterior probability is proportional to: –Prior probability –Likelihood of evidence Bayes’ Theorem assigns the appropriate weights in all cases: –With weak evidence, prior dominates results –With strong evidence, evidence dominates results Successive updating with several pieces of evidence gives same result as one-step updating with all the evidence

BAYES’ THEOREM – APPLICATION TO DATA ANALYSIS = A specific value for the frequency of some event E= Evidence related to that frequency = Prior probability that before observing evidence E = Posterior probability that after observing evidence E = Likelihood of observing evidence E if the true frequency is

LIKELIHOOD FUNCTION FOR DEMAND- BASED FAILURE RATES is the probability that a system or component will fail on each demand The likelihood of observing K failures in N demands is binomial; i.e.,

LIKLIHOOD FUNCTION FOR TIME-BASED FAILURE RATES is the frequency of failures per unit time The likelihood of observing K failures in T time units is Poisson; i.e.,

EXAMPLE OF BAYESIAN UPDATING PROBLEM 1:Use Bayes’ Theorem to calculate an updated (posterior) distribution for the component failure rate with evidence of 3 failures in 100 hours of operation PROBLEM 2:Repeat with evidence of 30 failures in 1,000 hours of operation Prior distribution of failure rate Failures per 100 hours

SOLUTION APPROACH 1.Discrete form of Bayes’ Theorem 2.Assume TRUE = CONSTANT 3.Likelihood function is Poisson

PROBLEM 1 RESULTS COMPARISON OF PRIOR AND POSTERIOR DISTRIBUTIONS Prior distribution Failures per 100 hours )(P j 0 Posterior distribution with E= Failures per 100 hours )(P j

PROBLEM 2 RESULTS COMPARISON OF PRIOR AND POSTERIOR DISTRIBUTIONS Prior distribution Failures per 100 hours )(P j 0 Posterior distribution with E= Failures per 100 hours )(P j.8

SOURCES OF UNCERTAINTY Interpretation and classification of events –Treatment of incipient failures –Extrapolation of data from other situations Determination of success data (number of demands and exposure/mission time) –Ability to detect failures –Run time versus latent time Sample size/statistical uncertainty –Possibility of missing data –Variability between components, vehicles, or missions Mathematical modeling –Assumption of constant failure rates

THE KNOWLEDGE ELICITATION PROCESS Motivate the experts – explain the importance of the assessment and the fact that information (not a firm prediction) is the goal Structure the interview – define the question to be answered, the units or scale for answering the question, and any inherent assumptions Exploratory discussions – discuss the parameter of interest to detect biases and induce the expert to reveal his or her true judgments Encoding – ask questions to encode the expert’s judgments, beginning with extreme percentiles Verification – construct the probability distribution and verify that the expert believes it is valid

ISSUES IN THE ELICITATION OF EXPERT OPINION Expert calibration Problem decomposition Structured versus unstructured group processes Mathematical aggregation of opinions versus the consensus approach

DEPENDENT FAILURES Functional dependencies –One system/component cannot operate because it depends on another system/component for its function Examples: –Lack of electric power fails system –False instrumentation signal means controller fails Treated in: –Event tree models –Split fraction models

DEPENDENT FAILURES Common cause failures –Shared cause leading to failure of two or more components within a short time interval (e.g., inherent defect in design, subtle environmental condition that is not explicitly identified) –Not failure of another component which can be explicitly modeled; e.g., breaker supplying power to both pumps Example –Lube oil filter plugging because of common defect and inattentive inspections Treated in –Split fraction models

DEPENDENT FAILURES Spatial dependencies –One system/component fails by virtue of close proximity to another system/component that has failed in a way that causes a cascading effect Examples –Shrapnel –Detonations –Fire Treated in –Event trees –Simulation models