Dynamic Bayesian Networks (DBNs)

Slides:



Advertisements
Similar presentations
CS188: Computational Models of Human Behavior
Advertisements

Exact Inference. Inference Basic task for inference: – Compute a posterior distribution for some query variables given some observed evidence – Sum out.
State Estimation and Kalman Filtering CS B659 Spring 2013 Kris Hauser.
Variational Methods for Graphical Models Micheal I. Jordan Zoubin Ghahramani Tommi S. Jaakkola Lawrence K. Saul Presented by: Afsaneh Shirazi.
CS498-EA Reasoning in AI Lecture #15 Instructor: Eyal Amir Fall Semester 2011.
Lauritzen-Spiegelhalter Algorithm
Exact Inference in Bayes Nets
Junction Trees And Belief Propagation. Junction Trees: Motivation What if we want to compute all marginals, not just one? Doing variable elimination for.
CSCI 121 Special Topics: Bayesian Networks Lecture #5: Dynamic Bayes Nets.
Lirong Xia Hidden Markov Models Tue, March 28, 2014.
Lirong Xia Approximate inference: Particle filter Tue, April 1, 2014.
Hidden Markov Models Reading: Russell and Norvig, Chapter 15, Sections
An Introduction to Variational Methods for Graphical Models.
Introduction of Probabilistic Reasoning and Bayesian Networks
Chapter 15 Probabilistic Reasoning over Time. Chapter 15, Sections 1-5 Outline Time and uncertainty Inference: ltering, prediction, smoothing Hidden Markov.
Overview of Inference Algorithms for Bayesian Networks Wei Sun, PhD Assistant Research Professor SEOR Dept. & C4I Center George Mason University, 2009.
Hidden Markov Models M. Vijay Venkatesh. Outline Introduction Graphical Model Parameterization Inference Summary.
Advanced Artificial Intelligence
1 Reasoning Under Uncertainty Over Time CS 486/686: Introduction to Artificial Intelligence Fall 2013.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
Probabilistic reasoning over time So far, we’ve mostly dealt with episodic environments –One exception: games with multiple moves In particular, the Bayesian.
… Hidden Markov Models Markov assumption: Transition model:
10/28 Temporal Probabilistic Models. Temporal (Sequential) Process A temporal process is the evolution of system state over time Often the system state.
1 Graphical Models in Data Assimilation Problems Alexander Ihler UC Irvine Collaborators: Sergey Kirshner Andrew Robertson Padhraic Smyth.
Learning with Bayesian Networks David Heckerman Presented by Colin Rickert.
Part 2 of 3: Bayesian Network and Dynamic Bayesian Network.
Graphical Models Lei Tang. Review of Graphical Models Directed Graph (DAG, Bayesian Network, Belief Network) Typically used to represent causal relationship.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
CPSC 322, Lecture 31Slide 1 Probability and Time: Markov Models Computer Science cpsc322, Lecture 31 (Textbook Chpt 6.5) March, 25, 2009.
11/14  Continuation of Time & Change in Probabilistic Reasoning Project 4 progress? Grade Anxiety? Make-up Class  On Monday?  On Wednesday?
Bayesian Networks Alan Ritter.
CS 188: Artificial Intelligence Fall 2009 Lecture 19: Hidden Markov Models 11/3/2009 Dan Klein – UC Berkeley.
Dynamic Bayesian Networks CSE 473. © Daniel S. Weld Topics Agency Problem Spaces Search Knowledge Representation Reinforcement Learning InferencePlanningLearning.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
Undirected Models: Markov Networks David Page, Fall 2009 CS 731: Advanced Methods in Artificial Intelligence, with Biomedical Applications.
Overview Particle filtering is a sequential Monte Carlo methodology in which the relevant probability distributions are iteratively estimated using the.
UIUC CS 498: Section EA Lecture #21 Reasoning in Artificial Intelligence Professor: Eyal Amir Fall Semester 2011 (Some slides from Kevin Murphy (UBC))
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 6.1: Bayes Filter Jürgen Sturm Technische Universität München.
CS Statistical Machine learning Lecture 24
CHAPTER 8 DISCRIMINATIVE CLASSIFIERS HIDDEN MARKOV MODELS.
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
CSC321: Neural Networks Lecture 16: Hidden Markov Models
QUIZ!!  In HMMs...  T/F:... the emissions are hidden. FALSE  T/F:... observations are independent given no evidence. FALSE  T/F:... each variable X.
Tractable Inference for Complex Stochastic Processes X. Boyen & D. Koller Presented by Shiau Hong Lim Partially based on slides by Boyen & Koller at UAI.
Probabilistic reasoning over time Ch. 15, 17. Probabilistic reasoning over time So far, we’ve mostly dealt with episodic environments –Exceptions: games.
Lecture 2: Statistical learning primer for biologists
1 Chapter 15 Probabilistic Reasoning over Time. 2 Outline Time and UncertaintyTime and Uncertainty Inference: Filtering, Prediction, SmoothingInference:
Bayesian networks and their application in circuit reliability estimation Erin Taylor.
Pattern Recognition and Machine Learning-Chapter 13: Sequential Data
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
CPSC 422, Lecture 17Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 17 Oct, 19, 2015 Slide Sources D. Koller, Stanford CS - Probabilistic.
CPS 170: Artificial Intelligence Markov processes and Hidden Markov Models (HMMs) Instructor: Vincent Conitzer.
Probability and Time. Overview  Modelling Evolving Worlds with Dynamic Baysian Networks  Simplifying Assumptions Stationary Processes, Markov Assumption.
1 Chapter 17 2 nd Part Making Complex Decisions --- Decision-theoretic Agent Design Xin Lu 11/04/2002.
Uncertain Observation Times Shaunak Chatterjee & Stuart Russell Computer Science Division University of California, Berkeley.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
Today Graphical Models Representing conditional dependence graphically
Probabilistic Reasoning Inference and Relational Bayesian Networks.
Instructor: Eyal Amir Grad TAs: Wen Pu, Yonatan Bisk Undergrad TAs: Sam Johnson, Nikhil Johri CS 440 / ECE 448 Introduction to Artificial Intelligence.
CS 541: Artificial Intelligence Lecture VIII: Temporal Probability Models.
CS 541: Artificial Intelligence Lecture VIII: Temporal Probability Models.
CS498-EA Reasoning in AI Lecture #23 Instructor: Eyal Amir Fall Semester 2011.
Today.
Probabilistic Reasoning over Time
CS 188: Artificial Intelligence Spring 2007
Inference Inference: calculating some useful quantity from a joint probability distribution Examples: Posterior probability: Most likely explanation: B.
Instructors: Fei Fang (This Lecture) and Dave Touretzky
Class #16 – Tuesday, October 26
Chapter14-cont..
Presentation transcript:

Dynamic Bayesian Networks (DBNs) Dave, Hsieh Ding Fei Frank, Yip Keung

Outline Introduction to DBNs Inference in DBNs Applications Conclusion Type of inference Exact inference Approximate inference Applications Conclusion

Introduction to DBNs Motivation Bayesian Network (BN) Models DBN Static nature of the problem domain Observable quantity is observed once for all Confidence in the observation is true for all time DBN Domains involving repeated observations Process dynamically evolves over time Examples: Monitoring a patient, traffic monitoring, etc.

Introduction to DBNs Assumptions The process is modeled as discrete time-slice At time 1, state is X(1) , at time t, state is X(t) P(X(1),…, X(t))=P(X(1)) P(X(1)|X(2) )…P(X(t)|X(1),…, X(t-1)) Markov property Given current state, the next state is independent of previous states P(X(1),…, X(t))=P(X(1)) P(X(1)|X(2) )…P(X(t)|X(t-1))

Introduction to DBNs DBN model (DAG representation) Edge means how tight the coupling is between nodes Effect is immediateedge within same time slice Effect is long termedge between time slices

Introduction to DBNs Special case of DBN  HMM State of HMM evolves in a Markovian way Model HMM as a simple DBN Each time slice contains two variables which are state q and observation o

Inference Type of Inference Prediction Monitoring Given a probability distribution over current state, predict the distribution over future states Monitoring Given the observation (evidence) in every time slice t, maintain the distribution over the current state Belief state at time T P(X(T) | o(1) ,…, o(T))

Inference Probability Estimation Explanation Given a sequence of observations in every time slice t, determine the distribution over each intermediate state P(X(t) | o(1) ,…, o(T)) for t = 1, 2, … , T Explanation Given an initial state and a sequence of observations o(1) ,…, o(T), determine the most likely sequence of states X(1) ,…, X(T)

Exact inference For most inference tasks, a belief state need to be maintained belief state A probability distribution over the current state This state summarize all information about history Need to be maintained compactly

Exact inference How to accomplish exact inference How to do this in a simple DBN  HMM Given a number of time slices, the DBN is just a very long BN with regular structure Standard Bayesian network algorithms can be used Probability estimation task Clique tree propagation algorithm Forward-backward algorithm

Exact inference Monitoring task Explanation task Prediction task Only the forward pass of forward-backward algorithm Explanation task Viterbi’s algorithm Prediction task Only base on the current belief state because it already have the history information

Exact inference dHugin : an exact inference computational system Inference method of classical discrete time-series analysis Allows discrete multivariate dynamic system

dHugin introduce notion of dynamic time window Contain several time slice and represent by junction tree Operations: window expansion and reduction Expand window to perform forecasting Inference are formulated in terms of message passing in junction tree

dHugin Window expansion Move k new consecutive time slices to the forecast model Move the k oldest time slices of the forecast model to the time window Moralize the compound graph including the graph in window and the new k slices Triangulate the time window Construct new junction tree

dHugin Window reduction Suppose has k+1 time slices in time window make the k oldest slices in time window become k backward smoothing models The remain (k+1)’st slices is the new time window

Forecasting Calculate estimates of the distributions of future variables given past observations and present variables Forecasting within window Propagation Forecasting beyond the window A series of alternating expansion and reduction step Propagation performed in each step

Problem of Exact inference Drawback: complex and require large space for computations Key issue is how to maintain the belief state Represent it naively Require an exponential number of entries Cannot represent it compactly by exploiting the structure no conditional independence structure Variables becomes correlated each other when time goes on Prevent using factorization ideas Not even conditionally independent within this time slice

Approximate Inference Objective Try to maintain and propagate an approximate belief state when the state space is very large in dynamic process It improves the complexity of probabilistic inference

Approximate Inference Two approaches Structural approximation Ignore weak correlations between variables in a belief state Stochastic simulation Randomly sample from the states in the belief state

Structural Approximation Problems in exact inference All variables in a belief state are correlated Belief state is expressed as full joint distribution  Need exponential number of table entries Objective of structural approximation Use factorization in order to represent complex system compactly by exploiting the fact that each variable has weak interaction with each other

Structural Approximation Example (monitor a freeway with multiple cars) States of different cars (e.g velocity,location..etc) become correlated after a certain period of time Approximation is to assume that the correlations are not very strong Each car can be considered as independent The approximate belief state can be represented in a factorized way, as a product of separate distributions, one for each car

Structural Approximation We can define a set of disjoint clusters Y1,…, Yk such that Y = Y1  Y2  …  Yk . We maintain an approximate belief state : If this approximate belief state of time t is simply propagated forward to time t+1, all variables would become correlated again

Structural Approximation It can be solved by executing the below process At each time t, we take and propagate it to time t+1, obtain a new distribution Approximate using independent marginal Compute for every I Ie. The product of each marginal is

Structural Approximation Two sources of error The accumulated error results from propagation The error results from approximation of Errors are bounded due to two opposing forces Propagation from time t to time t+1 adds noise to exact and approximate belief state  reduce difference between them  reduce error Approximation  increase error

Stochastic Simulation Likelihood Weighting (LW) Find the approximate belief state using sampling Algorithm of LW

Stochastic Simulation Drawback LW generates the samples at time t according to prior distribution (depends on condition of samples at time t-1) Observation affects the weights, but not the choice of samples Samples generated get increasingly irrelevant when time grows as some samples are not likely to happen to explain the current observation Example of monitoring car’s location

Stochastic Simulation Samples at t = 5 are more distributed, far away from exact location of vehicle An improved algorithm called Survival-Of-Fittest is used

Stochastic Simulation Survival-Of-Fittest (SOF) Propagate likely samples more often than unlikely samples Algorithm of SOF

Stochastic Simulation Belief state propagation over time (a) exact belief state (b) belief state by using LW (b) belief state by using SOF

Application Robot localization Track a robot moving around in an environment State variables x, y location Orientation Transition model corresponds to motion Next position is a Gaussian around a linear function of current position Observation model Probability that sonar detect an obstacle

Conclusion Concept DBNs Inference in DBNs Applications Four types of inference Exact inference dHugin Approximate inference Structural approximation Search –based Stochastic simulation Applications robot localization