September 28, 2000 Improved Simultaneous Data Reconciliation, Bias Detection and Identification Using Mixed Integer Optimization Methods Presented by:

Slides:



Advertisements
Similar presentations
COMM 472: Quantitative Analysis of Financial Decisions
Advertisements

Managerial Economics in a Global Economy
Computational Statistics. Basic ideas  Predict values that are hard to measure irl, by using co-variables (other properties from the same measurement.
Threshold Autoregressive. Several tests have been proposed for assessing the need for nonlinear modeling in time series analysis Some of these.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
A.M. Alonso, C. García-Martos, J. Rodríguez, M. J. Sánchez Seasonal dynamic factor model and bootstrap inference: Application to electricity market forecasting.
Hypothesis Testing Steps in Hypothesis Testing:
CHAPTER 2 Building Empirical Model. Basic Statistical Concepts Consider this situation: The tension bond strength of portland cement mortar is an important.
Use of Kalman filters in time and frequency analysis John Davis 1st May 2011.
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
Budapest May 27, 2008 Unifying mixed linear models and the MASH algorithm for breakpoint detection and correction Anders Grimvall, Sackmone Sirisack, Agne.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
1 Detection and Analysis of Impulse Point Sequences on Correlated Disturbance Phone G. Filaretov, A. Avshalumov Moscow Power Engineering Institute, Moscow.
Introduction Data and simula- tion methodology VaR models and estimation results Estimation perfor- mance analysis Conclusions Appendix Doctoral School.
Visual Recognition Tutorial
Linear Methods for Regression Dept. Computer Science & Engineering, Shanghai Jiao Tong University.
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
SYSTEMS Identification
Dynamic lot sizing and tool management in automated manufacturing systems M. Selim Aktürk, Siraceddin Önen presented by Zümbül Bulut.
Independent Component Analysis (ICA) and Factor Analysis (FA)
Lec 6, Ch.5, pp90-105: Statistics (Objectives) Understand basic principles of statistics through reading these pages, especially… Know well about the normal.
Prediction and model selection
Topic 3: Regression.
Estimation and the Kalman Filter David Johnson. The Mean of a Discrete Distribution “I have more legs than average”
Basics of regression analysis
Variance and covariance Sums of squares General linear models.
Adaptive Signal Processing
12 Autocorrelation Serial Correlation exists when errors are correlated across periods -One source of serial correlation is misspecification of the model.
Principles of the Global Positioning System Lecture 13 Prof. Thomas Herring Room A;
Algorithm Taxonomy Thus far we have focused on:
Introduction to Adaptive Digital Filters Algorithms
Empirical Financial Economics Asset pricing and Mean Variance Efficiency.
Oceanography 569 Oceanographic Data Analysis Laboratory Kathie Kelly Applied Physics Laboratory 515 Ben Hall IR Bldg class web site: faculty.washington.edu/kellyapl/classes/ocean569_.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
CHAPTER 4 Adaptive Tapped-delay-line Filters Using the Least Squares Adaptive Filtering.
EDGE DETECTION IN COMPUTER VISION SYSTEMS PRESENTATION BY : ATUL CHOPRA JUNE EE-6358 COMPUTER VISION UNIVERSITY OF TEXAS AT ARLINGTON.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
HQ U.S. Air Force Academy I n t e g r i t y - S e r v i c e - E x c e l l e n c e Improving the Performance of Out-of-Order Sigma-Point Kalman Filters.
The Group Lasso for Logistic Regression Lukas Meier, Sara van de Geer and Peter Bühlmann Presenter: Lu Ren ECE Dept., Duke University Sept. 19, 2008.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
PROCESS MODELLING AND MODEL ANALYSIS © CAPE Centre, The University of Queensland Hungarian Academy of Sciences Statistical Model Calibration and Validation.
Modern Navigation Thomas Herring MW 11:00-12:30 Room
Academic Research Academic Research Dr Kishor Bhanushali M
NCAF Manchester July 2000 Graham Hesketh Information Engineering Group Rolls-Royce Strategic Research Centre.
The Restricted Matched Filter for Distributed Detection Charles Sestok and Alan Oppenheim MIT DARPA SensIT PI Meeting Jan. 16, 2002.
Bootstrap Event Study Tests Peter Westfall ISQS Dept. Joint work with Scott Hein, Finance.
Simulation Study for Longitudinal Data with Nonignorable Missing Data Rong Liu, PhD Candidate Dr. Ramakrishnan, Advisor Department of Biostatistics Virginia.
Comp. Genomics Recitation 10 4/7/09 Differential expression detection.
An Introduction To The Kalman Filter By, Santhosh Kumar.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Normal Equations The Orthogonality Principle Solution of the Normal Equations.
Dynamic Models, Autocorrelation and Forecasting ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
CY3A2 System identification Input signals Signals need to be realisable, and excite the typical modes of the system. Ideally the signal should be persistent.
5.3 Mixed Integer Nonlinear Programming Models. A Typical MINLP Model.
There is a hypothesis about dependent and independent variables The relation is supposed to be linear We have a hypothesis about the distribution of errors.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Kalman Filter with Process Noise Gauss- Markov.
Computacion Inteligente Least-Square Methods for System Identification.
Multivariate statistical methods. Multivariate methods multivariate dataset – group of n objects, m variables (as a rule n>m, if possible). confirmation.
CWR 6536 Stochastic Subsurface Hydrology Optimal Estimation of Hydrologic Parameters.
Chapter 7. Classification and Prediction
Linear Regression.
Kakhramon Yusupov June 15th, :30pm – 3:00pm Session 3
5.3 Mixed-Integer Nonlinear Programming (MINLP) Models
5.3 Mixed Integer Nonlinear Programming Models
Autocorrelation.
Multivariate Methods Berlin Chen
Multivariate Methods Berlin Chen, 2005 References:
Autocorrelation.
Threshold Autoregressive
Presentation transcript:

September 28, 2000 Improved Simultaneous Data Reconciliation, Bias Detection and Identification Using Mixed Integer Optimization Methods Presented by: Tyler A. Soderstrom

Presentation Overview Background MILP Method Extension to Nonlinear Problems Inclusion of Statistical Tests as Constraints Multiple System Models as Constraints Correlated Data Conclusions

Background Data Reconciliation – –Optimal estimates for noisy measurements Bias Detection / Identification – –Determine presence and location of bias Problems are closely related –Presence of bias skews reconciliation results –Common techniques require reconciliation residuals

MILP Method Type of systems considered – Process data matrix –

MILP Method Problem Formulation –

MILP Method Realizable form –

MILP Method Bias constraint region l  ll U  l U  ll U  l U

Tuning Issues Horizon Length Binary variable weighting Bias bounding and thresholding

Extension to Nonlinear Problems Straightforward constraint replacement – with Modified problem is a MINLP –Tougher class of problem –Solution technology is not as mature –Global solution is not guaranteed

Solution Methods Outer Approximation / Equality Relaxation –DICOPT (J. Viswanathan and I. E. Grossman) Random Search –Genetic Algorithm (J. Holland) Meta-Heuristics –Tabu Search (F. Glover)

Basic Method Extensions Make use of Past Information Formulate Extensions as Additional Problem Constraints –Incorporate Common Statistical Tests –Include Empirical Data Model –Compensate for Non-Ideal Process Data May Require Modifications to Objective

Using Previous Estimates Moving Horizon Estimation Problem Previous Estimates of Process Variables and Biases Can be Made Available Including Past Information In Current Problem Execution Improves Stability and Convergence of Estimates Objective is Pathway to Past Information

Objective Modifications Add to Objective Φ Where and are Estimates from Previous Execution Convert to Realizable Form

Realizable Form Objective Φ Additional Constraints

Bias Penalty Term Inclusion Depends on Objectives –Most Will be Zero –May Delay Identification of Bias –If Bias is Persistent, Improve Estimate Convergence Not Important for Optimization Engine –Previous Solution Warm Starts Current Run

Statistical Tests as Constraints Tests are based on Hypothesis Testing –Test Statistic is Proposed –Statistic is Calculated with Current Data –If Statistic Exceeds a Threshold Value (Related to Level of Confidence) Bias is Present Statistic is defined as a problem variable Definition added as a problem constraint Constraint bounding Statistic below threshold Forces the no bias conclusion at solution

Mathematical Description Hypothesis Testing –H 0 there is no bias in the process data –H 1 there is at least 1 bias in the process data –Choice Depends on value of test statistic at a given level of significance Test StatisticZ Add the Following Constraints to Problem Definition Constraint Z   h(y) Null Enforcement Constraint |Z  < Z 

Objective Modifications Previous Description may be Infeasible Define new Constraint Violation Variable and Change Constraint Null Enforcement Constraint: |Z|  < Z   Penalize Variable in Objective Objective:  =  old + w i 

MILP Example

Embedded PC Test Principal Component Test –Form Matrix s.t. it contains eigenvectors of –y e contains principle component scores –scores are normal zero mean, unit variance

Embedded PC Test Principal Component Test – Threshold Value at confidence level  –Perform Test on Averaged Measurements Enhance Power of Test Single set of additional constraints is used in the formation of

Embedded PC Test Additional Constraints

Performance Measures Average number of Type I Errors Overall Power

Simulation Results

Discussion of Results No Benefit to OP –Estimates from basic method usually pass tests without specific enforcement Test Enforcement Increases AVTI –Other biases forced to become active to lower statistics Usually Global Statistical Tests Used First –require nonlinear equations

Non-Ideal Data Compensation Serially Correlated Data Requires New Measurement Noise Model Error Sequence,, Forms a Stationary Process Assume no Cross Correlation

Statistical Tests When Data are Serially Correlated Tests on Individual Vectors are Unaffected Statistical Tests are Often Used on Vectors of Averaged Measurement –Increase Power of the Test –Autocorrelation Invalidates Test Assumptions –Procedure Must be Modified

Statistical Tests When Data are Serially Correlated Most Test Statistics Require Covariance of N Averaged Measurements: –Time Independent Data: –Autocorrelated Data:

Methods of Dealing With Serially Correlated Data Correcting the Variance –Requires all Autocorrelation Coefficients –Coefficients can be calculated analytically if noise model is known (e.g. time series) –Otherwise coefficients can be estimated Prewhitening –Filtering Approach –Requires expression of noise model

Methods of Dealing With Serially Correlated Data Prewhitening (cont.) –Calculate a “approximately independent” sequence –Apply tests designed for independent data

Implementing Compensation Within MIP Framework Correcting the Variance –Unknown Correlation Model Estimate with bias free data Include, calculated using as parameters in MIP program –Known Correlation Model calculated analytically are used as parameters in MIP program Include Modified Test as Set of Constraints

Implementing Compensation Within MIP Framework Prewhitening –Uncorrelated residuals written as functions of noise model parameters and measurements –Equations included as constraints –Tests on uncorrelated residuals included as constraints

Conclusions MIP bias detection / identification performs better than several other methods High Power / Low Occurrence of False Identification Straightforward implementation Method Enhancements –Consider Past Information –Include Statistical Tests in Constraints Univariate Tests Do Not Improve Performance Global Tests May Help, but require nonlinear equations –Handling Autocorrelated Data

Future Work Investigate Additional Constraints with Nonlinear Models –Nonlinear Statistical Tests –Improve sensitivity to small bias Compare solution methods on larger nonlinear models Extend Method to Dynamic Models –Discrete vs. Continuous –Linear and Nonlinear –Computational Issues