Pat Langley Dileep George Stephen Bay Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford,

Slides:



Advertisements
Similar presentations
Computational Revision of Ecological Process Models
Advertisements

Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Pat Langley Institute for the Study of Learning and Expertise Javier Sanchez CSLI / Stanford University Ljupco Todorovski Saso Dzeroski Jozef Stefan Institute.
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Pat Langley Center for the Study of Language and Information Stanford University, Stanford, California
Filtering Information in Complex Temporal Domains
Pat Langley School of Computing and Informatics Arizona State University Tempe, Arizona Institute for the Study of Learning and Expertise Palo Alto, California.
Pat Langley Institute for the Study of Learning and Expertise Palo Alto, California and Center for the Study of Language and Information Stanford University,
Pat Langley School of Computing and Informatics Arizona State University Tempe, Arizona Computational Discovery.
Pat Langley Center for the Study of Language and Information Stanford University, Stanford, California
Pat Langley Center for the Study of Language and Information Stanford University, Stanford, California
Pat Langley Institute for the Study of Learning and Expertise 2164 Staunton Court, Palo Alto, California and School of Computing and Informatics Arizona.
Pat Langley School of Computing and Informatics Arizona State University Tempe, Arizona Computational Discovery of Explanatory Process Models Thanks to.
Javier Sanchez Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California.
Pat Langley Institute for the Study of Learning and Expertise Palo Alto, California and Center for the Study of Language and Information Stanford University,
Pat Langley Computer Science and Engineering / Psychology Arizona State University Tempe, Arizona Challenges and Opportunities in Informatics Research.
The Robert Gordon University School of Engineering Dr. Mohamed Amish
Design of Experiments Lecture I
Fast Algorithms For Hierarchical Range Histogram Constructions
A.M. Alonso, C. García-Martos, J. Rodríguez, M. J. Sánchez Seasonal dynamic factor model and bootstrap inference: Application to electricity market forecasting.
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
Learning Process-Based Models of Dynamic Systems Nikola Simidjievski Jozef Stefan Institute, Slovenia HIPEAC 2014 LJUBLJANA.
Pat Langley School of Computing and Informatics Arizona State University Tempe, Arizona Institute for the Study of Learning and Expertise Palo Alto, California.
Formal Methods in Software Engineering Credit Hours: 3+0 By: Qaisar Javaid Assistant Professor Formal Methods in Software Engineering1.
Lecture 8 Relationships between Scale variables: Regression Analysis
Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.
A Framework for Discovering Anomalous Regimes in Multivariate Time-Series Data with Local Models Stephen Bay Stanford University, and Institute for the.
Relational Data Mining in Finance Haonan Zhang CFWin /04/2003.
Knowledge Acquisitioning. Definition The transfer and transformation of potential problem solving expertise from some knowledge source to a program.
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Chapter Two SCIENTIFIC METHODS IN BUSINESS
P ROCESSES AND C ONSTRAINTS IN S CIENTIFIC M ODEL C ONSTRUCTION Will Bridewell † and Pat Langley †‡ † Cognitive Systems Laboratory, CSLI, Stanford University.
MATH408: Probability & Statistics Summer 1999 WEEKS 8 & 9 Dr. Srinivas R. Chakravarthy Professor of Mathematics and Statistics Kettering University (GMI.
Applying Multi-Criteria Optimisation to Develop Cognitive Models Peter Lane University of Hertfordshire Fernand Gobet Brunel University.
IE-331: Industrial Engineering Statistics II Spring 2000 WEEK 1 Dr. Srinivas R. Chakravarthy Professor of Operations Research and Statistics Kettering.
CSC2535: 2013 Advanced Machine Learning Lecture 3a: The Origin of Variational Bayes Geoffrey Hinton.
A prediction or a forecast is a statement about the way things will happen in the future, often but not always based on experience or knowledge. While.
Framework for K-12 Science Education
Research Methods in Nursing. Examining History 1600’s is the age of reasoning; finding reason and experimenting what is observed. Isaac Newton is a pioneer.
Virginia Standard of Learning BIO.1a-m
Unit 2: Engineering Design Process
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
The student will demonstrate an understanding of how scientific inquiry and technological design, including mathematical analysis, can be used appropriately.
Discovering Dynamic Models Lecture 21. Dynamic Models: Introduction Dynamic models can describe how variables change over time or explain variation by.
Pat Langley Adam Arvay Department of Computer Science University of Auckland Auckland, NZ Heuristic Induction of Rate-Based Process Models Thanks to W.
ENM 503 Lesson 1 – Methods and Models The why’s, how’s, and what’s of mathematical modeling A model is a representation in mathematical terms of some real.
Remarks: 1.When Newton’s method is implemented has second order information while Gauss-Newton use only first order information. 2.The only differences.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Developing and Evaluating Theories of Behavior.
WHAT IS THE NATURE OF SCIENCE?. SCIENTIFIC WORLD VIEW 1.The Universe Is Understandable. 2.The Universe Is a Vast Single System In Which the Basic Rules.
Estimating Component Availability by Dempster-Shafer Belief Networks Estimating Component Availability by Dempster-Shafer Belief Networks Lan Guo Lane.
Learning and Acting with Bayes Nets Chapter 20.. Page 2 === A Network and a Training Data.
Data Mining and Decision Support
Nonlinear differential equation model for quantification of transcriptional regulation applied to microarray data of Saccharomyces cerevisiae Vu, T. T.,
A Framework to Predict the Quality of Answers with Non-Textual Features Jiwoon Jeon, W. Bruce Croft(University of Massachusetts-Amherst) Joon Ho Lee (Soongsil.
Pat Langley School of Computing and Informatics Arizona State University Tempe, Arizona Computational Assistance for Systems Biology of Aging Thanks to.
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, CA
Paradigms of Knowing in Communication Research. Paradigms in social science provide a viewpoint or set of assumptions that frame the research process.
Introduction It had its early roots in World War II and is flourishing in business and industry with the aid of computer.
Borrett et al Computational Discovery of Process Models for Aquatic Ecosystems August 2006 Ecological Society of America, Memphis, TN Natasa Atanasova.
Remember the equation of a line: Basic Linear Regression As scientists, we find it an irresistible temptation to put a straight line though something that.
Operations Research Chapter one.
C.-S. Shieh, EC, KUAS, Taiwan
Chapter 6 Calibration and Application Process
Developing and Evaluating Theories of Behavior
Regression Models - Introduction
Causal Models Lecture 12.
Observation Use of one or more of the senses to gather information
Presented By: Darlene Banta
Presentation transcript:

Pat Langley Dileep George Stephen Bay Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, CA Kazumi Saito NTT Communication Science Laboratories Soraku, Kyoto, JAPAN Robust Induction of Process Models from Time-Series Data This research was funded in part by NTT Communication Science Laboratories and in part by Grant NCC from NASA Ames Research Center.

A Process Model for an Aquatic Ecosystem model AquaticEcosystem variables: phyto, zoo, nitro, residue observables: phyto, nitro process phyto_exponential_decay equations:d[phyto,t,1] = phyto equations:d[phyto,t,1] = phyto d[residue,t,1] = phyto process zoo_exponential_decay equations:d[zoo,t,1] = zoo equations:d[zoo,t,1] = zoo d[residue,t,1] = process zoo_phyto_predation equations:d[zoo,t,1] = zoo equations:d[zoo,t,1] = zoo d[residue,t,1] = zoo d[phyto,t,1] = zoo process nitro_uptake conditions:nitro > 0 conditions:nitro > 0 equations:d[phyto,t,1] = phyto equations:d[phyto,t,1] = phyto d[nitro,t,1] = phyto process nitro_remineralization; equations:d[nitro,t,1] = residue equations:d[nitro,t,1] = residue d[residue,t,1 ] = residue

Predictions from the Ecosystem Model

Advantages of Quantitative Process Models they refer to notations and mechanisms familiar to scientists; they refer to notations and mechanisms familiar to scientists; they embed quantitative relations within qualitative structure; they embed quantitative relations within qualitative structure; they provide dynamical predictions of changes over time; they provide dynamical predictions of changes over time; they offer causal and explanatory accounts of phenomena; they offer causal and explanatory accounts of phenomena; while retaining the modularity needed to support induction. while retaining the modularity needed to support induction. Process models are a good target for discovery systems because: Quantitative process models provide an important alternative to formalisms used currently in machine learning and discovery.

Observed values for a set of continuous variables as they vary over time or situations Generic processes that characterize causal relationships among variables in terms of conditional equations Inductive Process Modeling A specific process model that explains the observed values and predicts future data accurately Induction training data background knowledge learned model

Generic Processes as Background Knowledge the variables involved in a process and their types; the variables involved in a process and their types; the parameters appearing in a process and their ranges; the parameters appearing in a process and their ranges; the forms of conditions on the process; and the forms of conditions on the process; and the forms of associated equations and their parameters. the forms of associated equations and their parameters. Our framework casts background knowledge as generic processes that specify: Generic processes are building blocks from which one can compose a specific quantitative process model.

Generic Processes for Aquatic Ecosystems generic process exponential_decaygeneric process remineralization variables: S{species}, D{detritus} variables: N{nutrient}, D{detritus} variables: S{species}, D{detritus} variables: N{nutrient}, D{detritus} parameters: [0, 1] parameters: [0, 1] parameters: [0, 1] parameters: [0, 1] equations:d[S,t,1] = 1 S equations:d[N, t,1] = D equations:d[S,t,1] = 1 S equations:d[N, t,1] = D d[D,t,1] = Sd[D, t,1] = 1 D generic process predationgeneric process constant_inflow variables: S1{species}, S2{species}, D{detritus} variables: N{nutrient} variables: S1{species}, S2{species}, D{detritus} variables: N{nutrient} parameters: [0, 1], [0, 1] parameters: [0, 1] parameters: [0, 1], [0, 1] parameters: [0, 1] equations:d[S1,t,1] = S1 equations:d[N,t,1] = equations:d[S1,t,1] = S1 equations:d[N,t,1] = d[D,t,1] = (1 ) S1 d[S2,t,1] = 1 S1 generic process nutrient_uptake variables: S{species}, N{nutrient} variables: S{species}, N{nutrient} parameters: [0, ], [0, 1], [0, 1] parameters: [0, ], [0, 1], [0, 1] conditions:N > conditions:N > equations:d[S,t,1] = S equations:d[S,t,1] = S d[N,t,1] = 1 S

Previous Results: The IPM Algorithm 1. Find all ways to instantiate known generic processes with specific variables, subject to type constraints; 2. Combine instantiated processes into candidate generic models, with limits on the total number of processes; 3. For each generic model, carry out gradient descent search through parameter space to find good parameter values; 4. Select the parameterized model that produces the lowest mean squared error on the training data. Langley et al. (2002) reported IPM, an algorithm that constructs process models from generic components in four stages: We showed that IPM could induce accurate process models from noisy time series, but it tended to include extra processes.

The Revised IPM Algorithm Accepts as input those variables that can appear in the induced model, both observable and unobservable; Accepts as input those variables that can appear in the induced model, both observable and unobservable; Utilizes the parameter-fitting routine to estimate initial values for unobservable variables; Utilizes the parameter-fitting routine to estimate initial values for unobservable variables; Invokes the parameter-fitting method to induce the thresholds on process conditions; and Invokes the parameter-fitting method to induce the thresholds on process conditions; and Selects the parameterized model with the lowest description length: M d = (M v + M c ) log (n) + n log (M e ). Selects the parameterized model with the lowest description length: M d = (M v + M c ) log (n) + n log (M e ). We have revised and extended the IPM algorithm so that it now: We have evaluated the new system on synthetic and natural data.

Evaluation of the IPM Algorithm 1. We used the aquatic ecosystem model to generate data sets over 100 time steps for the variables nitro and phyto; 2. We replaced each true value x with x (1 + r n), where r followed a Gaussian distribution ( = 0, = 1) and n > 0; 3. We ran IPM on these noisy data, giving it type constraints and generic processes as background knowledge. To demonstrate IPM's ability to induce process models, we ran it on synthetic data for a known system: In two experiments, we let IPM determine the initial values and thresholds given the correct structure; in a third study, we let it search through a space of 256 generic model structures.

Experimental Results with IPM The main results of our studies with IPM on synthetic data were: 1. The system infers accurate estimates for the initial values of unobservable variables like zoo and residue; 2. The system induces estimates of condition thresholds on nitro that are close to the target values; and 3. The MDL criterion selects the correct model structure in all runs with 5% noise, but only 40% of runs with 10% noise. These suggest that the basic approach is sound, but that we should consider other MDL schemes and other responses to overfitting.

Results with Unobserved Initial Values

Electric Power on the International Space Station

Telemetry Data from Space Station Batteries Predictor variables included the batterys current and temperature.

Induced Process Model for Battery Behavior model Battery variables: Rs, Vcb, soc, Vt, i, temperature observable: soc, Vt, i, temperature process voltage_chargeprocess voltage_discharge conditions:i 0 conditions:i < 0 conditions:i 0 conditions:i < 0 equations:Vt = Vcb Rs i equations:Vt = Vcb 1.0 / (Rs + 1.0) equations:Vt = Vcb Rs i equations:Vt = Vcb 1.0 / (Rs + 1.0) process charge_transfer equations:d[soc,t,1] = i Vcb/ equations:d[soc,t,1] = i Vcb/ process quadratic_influence_Vcb_soc equations:Vcb = soc soc equations:Vcb = soc soc process linear_influence_Vcb_temp equations:Vcb = temperature equations:Vcb = temperature process linear_influence_Rs_soc equations:Rs = soc equations:Rs = soc

Results on Battery Test Data

Best Fit to Data on Protozoan Predation

computational scientific discovery (e.g., Langley et al., 1983) computational scientific discovery (e.g., Langley et al., 1983) knowledge-based learning methods (e.g., ILP, theory revision) knowledge-based learning methods (e.g., ILP, theory revision) qualitative physics and simulation (e.g., Forbus, 1984) qualitative physics and simulation (e.g., Forbus, 1984) scientific simulation environments (e.g., STELLA, MATLAB ) scientific simulation environments (e.g., STELLA, MATLAB ) Intellectual Influences Our work on inductive process modeling incorporates ideas from many traditions: However, the most similar research comes from Todorovski and Dzeroski (1997) and from Bradley, Easley, and Stolle (2001). Their approaches also use knowledge to guide the induction of differential equation models, though without a process formalism.

Directions for Future Research produce additional results on other scientific data sets produce additional results on other scientific data sets develop more robust methods for fitting model parameters develop more robust methods for fitting model parameters explore alternative techniques that mitigate overfitting explore alternative techniques that mitigate overfitting extend the approach to handle data sets with missing values extend the approach to handle data sets with missing values implement heuristic methods for searching the model space implement heuristic methods for searching the model space utilize knowledge of subsystems to further constrain search utilize knowledge of subsystems to further constrain search Despite our progress to date, we need further work in order to: Our goal is a robust approach to inductive process modeling that can aid scientists and engineers in model construction.

End of Presentation