UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering On-line Alert Systems for Production Plants A Conflict Based Approach.

Slides:



Advertisements
Similar presentations
Richard M. Jacobs, OSA, Ph.D.
Advertisements

A Fast Estimation of SRAM Failure Rate Using Probability Collectives Fang Gong Electrical Engineering Department, UCLA Collaborators:
ASSESSING RESPONSIVENESS OF HEALTH MEASUREMENTS. Link validity & reliability testing to purpose of the measure Some examples: In a diagnostic instrument,
June 15, 2005 Integration of Diagnostics and Prognostics.
Hypothesis Testing Steps in Hypothesis Testing:
The Essentials of 2-Level Design of Experiments Part I: The Essentials of Full Factorial Designs The Essentials of 2-Level Design of Experiments Part I:
G. Alonso, D. Kossmann Systems Group
STAT 497 APPLIED TIME SERIES ANALYSIS
Topic 6: Introduction to Hypothesis Testing
1 1 Slide © 2009, Econ-2030 Applied Statistics-Dr Tadesse Chapter 10: Comparisons Involving Means n Introduction to Analysis of Variance n Analysis of.
UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering CSCE 580 Artificial Intelligence Ch.5 [P]: Propositions and Inference Sections.
Hidden Markov Model Special case of Dynamic Bayesian network Single (hidden) state variable Single (observed) observation variable Transition probability.
1 EE 616 Computer Aided Analysis of Electronic Networks Lecture 9 Instructor: Dr. J. A. Starzyk, Professor School of EECS Ohio University Athens, OH,
Analysis of Variance Chapter 3Design & Analysis of Experiments 7E 2009 Montgomery 1.
Chapter 11 Multiple Regression.
Meta-Analysis and Strategy Research Dan R. Dalton Kelley School of Business Indiana University.
Business Statistics - QBM117 Interval estimation for the slope and y-intercept Hypothesis tests for regression.
Regression Model Building Setting: Possibly a large set of predictor variables (including interactions). Goal: Fit a parsimonious model that explains variation.
Introduction to Regression Analysis, Chapter 13,
Analysis of Simulation Results Andy Wang CIS Computer Systems Performance Analysis.
V. Rouillard  Introduction to measurement and statistical analysis ASSESSING EXPERIMENTAL DATA : ERRORS Remember: no measurement is perfect – errors.
AM Recitation 2/10/11.
Water Contamination Detection – Methodology and Empirical Results IPN-ISRAEL WATER WEEK (I 2 W 2 ) Eyal Brill Holon institute of Technology, Faculty of.
HAWKES LEARNING SYSTEMS math courseware specialists Copyright © 2010 by Hawkes Learning Systems/Quant Systems, Inc. All rights reserved. Chapter 14 Analysis.
Multiple testing correction
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 6 – Multiple comparisons, non-normality, outliers Marshall.
Simple Linear Regression
1 1 Slide © 2006 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Correlation.
Topic 6.1 Statistical Analysis. Lesson 1: Mean and Range.
Verification & Validation
Montecarlo Simulation LAB NOV ECON Montecarlo Simulations Monte Carlo simulation is a method of analysis based on artificially recreating.
Introduction to Statistical Quality Control, 4th Edition
1 Chapter 13 Analysis of Variance. 2 Chapter Outline  An introduction to experimental design and analysis of variance  Analysis of Variance and the.
LECTURER PROF.Dr. DEMIR BAYKA AUTOMOTIVE ENGINEERING LABORATORY I.
Ch5. Probability Densities II Dr. Deshi Ye
Chapter 16 The Chi-Square Statistic
Managerial Economics Demand Estimation & Forecasting.
A Process Control Screen for Multiple Stream Processes An Operator Friendly Approach Richard E. Clark Process & Product Analysis.
Chapter 8 Sampling Variability and Sampling Distributions.
Section 10.1 Confidence Intervals
5.1 Chapter 5 Inference in the Simple Regression Model In this chapter we study how to construct confidence intervals and how to conduct hypothesis tests.
Chapter 14 Repeated Measures and Two Factor Analysis of Variance
Sample Size Determination Text, Section 3-7, pg. 101 FAQ in designed experiments (what’s the number of replicates to run?) Answer depends on lots of things;
1 OUTPUT ANALYSIS FOR SIMULATIONS. 2 Introduction Analysis of One System Terminating vs. Steady-State Simulations Analysis of Terminating Simulations.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
T Test for Two Independent Samples. t test for two independent samples Basic Assumptions Independent samples are not paired with other observations Null.
Chapter 15 The Chi-Square Statistic: Tests for Goodness of Fit and Independence PowerPoint Lecture Slides Essentials of Statistics for the Behavioral.
Educational Research: Data analysis and interpretation – 1 Descriptive statistics EDU 8603 Educational Research Richard M. Jacobs, OSA, Ph.D.
Chapter 9: Introduction to the t statistic. The t Statistic The t statistic allows researchers to use sample data to test hypotheses about an unknown.
10 March 2016Materi ke-3 Lecture 3 Statistical Process Control Using Control Charts.
Statistical principles: the normal distribution and methods of testing Or, “Explaining the arrangement of things”
Copyright © 2009 Pearson Education, Inc t LEARNING GOAL Understand when it is appropriate to use the Student t distribution rather than the normal.
Anomaly Detection Carolina Ruiz Department of Computer Science WPI Slides based on Chapter 10 of “Introduction to Data Mining” textbook by Tan, Steinbach,
Bias Management in Time Changing Data Streams We assume data is generated randomly according to a stationary distribution. Data comes in the form of streams.
Chapter 14 Repeated Measures and Two Factor Analysis of Variance PowerPoint Lecture Slides Essentials of Statistics for the Behavioral Sciences Seventh.
Chapter 9 Introduction to the t Statistic
More on Inference.
Online Conditional Outlier Detection in Nonstationary Time Series
Chapter 8: Inference for Proportions
Inferences About Means from Two Groups
Towson University - J. Jung
More on Inference.
Inferences and Conclusions from Data
CSCI 5822 Probabilistic Models of Human and Machine Learning
Two-way analysis of variance (ANOVA)
The Essentials of 2-Level Design of Experiments Part I: The Essentials of Full Factorial Designs Developed by Don Edwards, John Grego and James Lynch.
Chapter 10 – Part II Analysis of Variance
Simulation Berlin Chen
Presentation transcript:

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering On-line Alert Systems for Production Plants A Conflict Based Approach

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Alert Systems Based on sensor readings, raises a flag in case of an abnormal situation Researchers construct a system using Bayesian Networks to detect abnormal situations and generate alerts

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Typical Production Plant S1S1 S2S2 SnSn C1C1 C2C2 CnCn

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Common Issues Engineers’ knowledge of plant is not sufficient for providing a causal structure Process is too complex to specify the possible faults and how to detect them based on sensor readings Difficult to determine the delay from event to effect Faults are so rare that statistics cannot be used to learn neither the structure nor the parameters of a model of the faults Difference between a true value and its sensor reading; true values should appear as hidden variables

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causal Approach Collect as much causal structure as possible and combine with a data driven learning method Limitations –Learning algorithms cannot cope with domains with a massive set of hidden variables –It is not obvious how such a model could be used for classifying abnormal behavior

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Proposed Method Learn a Bayesian Network representing normal operation only Does not require information about possible faults or modeling of abnormal behavior Faults are detected by measuring conflict between model and sensor readings

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Proposed Method Consists of two steps: –Learning a model of the sensors for normal operation –Using the learned model to monitor the system, initiate alerts and perform on-line diagnostics

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Learning a Model Can be done in many different ways In this paper the researchers analyzed the database of sensor readings during normal operation where the variables are the sensors and hidden variables are the components of the system

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Initiation of Alerts Sensor readings are received in a constant flow Readings are chopped up into time steps of 1 second Therefore every second we have evidence for every variable in the model

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Conflict Measure of Evidence Let evidence be e = {e 1,…,e n } and the conflict measure of the evidence is:

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Detecting Conflict In general during normal operating we expect: conf(e) ≤ 0 An indication of an abnormal situation is detected when: conf(e) > 0

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Problem with Conflict Measure A negative conflict value does not necessarily imply that we have a normal situation If sensors are strongly correlated during normal operation, the conflict level will be very negative A few conflicting sensors therefore will not cause the entire conflict to be positive To detect watch for jumps in the conflict level

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Tracing Source of Alert Greedy Conflict Resolution –Recursively remove the sensor reading that reduces the conflict the most –Stop when conflict is below a predefined threshold –Can be performed very fast using: fast retraction, lazy propagation or arithmetic circuits

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Conflict Resolution

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Coal Power Plant Network

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Sample Data set Used sample data from normal operation to learn model for coal mill Two data sets covered actual errors/abnormal situations: –The fall-pipe leading coal into the power plant becomes clogged –A temperature sensor becomes faulty

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Clogged Fall-Pipe Data

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Clogged Fall-Pipe Conflict Resolution Indicated that the sensor measuring the water- percentage in the coal can explain all the conflicts There is no sensor for Fall-Pipe clog detection so this is the closest sensor that could explain the conflict Result was consistent with the analysis of the plant Engineers

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Faulty-sensor Data

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Faulty Sensor Conflict Resolution Indicated that six significant sensors could explain the conflict Engineers indicated that four of the six were actually significant, the other two were anomalies due to the model

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Oil Production Facility Simulated cases for normal system operation Two other data sets covered errors/abnormal situations: –Simulated faults in the pumping system –Simulated faults in the cooling system

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Pump and Cooling Data

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Pump Data

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Cooling Data

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Change Point Detection Note that in the previous plots that the conflict measures where all negative indicating no faults However, remember the problem if the sensors are strongly correlated (slide 15) In order to perform conflict resolution the conflict threshold should be based on values observed during normal operation In order to detect changes in system operation need to track jumps in the conflict measure

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Change Point Detection Assume conflict values for normal operation are independent samples from normal distribution with fixed mean and variance Model the l’th conflict value as random variable with normal distribution f where mean  l and variance  l 2 are estimated using the last m observations:

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Change Point Detection To detect change point calculate the logarithmic loss for the last n observations and raise alert in case the value is above a predefined threshold:

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Change Point Detection This approach is sensitive to fluctuations and results in false positives; also has difficulty detecting drifts in conflict measure To alleviate this compare model with another model f’ where mean and variance estimated by shifting s observations back:

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Change Point Detection Compare models (score difference):

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Change Point Detection For normal operation score should be within the interval [-  :  ]  will determine the ratio of false positives and false negatives n determines the response time of the system m determines the relevant history s the number of observations to shift back

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Clogged Fall-Pipe Data (n=m=5, s=20)

UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Clogged Fall-Pipe Data (n=m=10, s=40)