Dynamic Bayesian Networks and Particle Filtering COMPSCI 276 (chapter 15, Russel and Norvig) 2007.

Slides:



Advertisements
Similar presentations
CSCE643: Computer Vision Bayesian Tracking & Particle Filtering Jinxiang Chai Some slides from Stephen Roth.
Advertisements

Dynamic Bayesian Networks (DBNs)
Lirong Xia Approximate inference: Particle filter Tue, April 1, 2014.
Rao-Blackwellised Particle Filtering Based on Rao-Blackwellised Particle Filtering for Dynamic Bayesian Networks by Arnaud Doucet, Nando de Freitas, Kevin.
Visual Tracking CMPUT 615 Nilanjan Ray. What is Visual Tracking Following objects through image sequences or videos Sometimes we need to track a single.
Chapter 15 Probabilistic Reasoning over Time. Chapter 15, Sections 1-5 Outline Time and uncertainty Inference: ltering, prediction, smoothing Hidden Markov.
Introduction to Sampling based inference and MCMC Ata Kaban School of Computer Science The University of Birmingham.
Probabilistic reasoning over time So far, we’ve mostly dealt with episodic environments –One exception: games with multiple moves In particular, the Bayesian.
… Hidden Markov Models Markov assumption: Transition model:
1 Graphical Models in Data Assimilation Problems Alexander Ihler UC Irvine Collaborators: Sergey Kirshner Andrew Robertson Padhraic Smyth.
CS 188: Artificial Intelligence Fall 2009 Lecture 20: Particle Filtering 11/5/2009 Dan Klein – UC Berkeley TexPoint fonts used in EMF. Read the TexPoint.
Particle Filters Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics TexPoint fonts used in EMF. Read.
Graphical Models for Mobile Robot Localization Shuang Wu.
Machine Learning CUNY Graduate Center Lecture 7b: Sampling.
SLAM: Simultaneous Localization and Mapping: Part I Chang Young Kim These slides are based on: Probabilistic Robotics, S. Thrun, W. Burgard, D. Fox, MIT.
Temporal Processes Eran Segal Weizmann Institute.
Nonlinear and Non-Gaussian Estimation with A Focus on Particle Filters Prasanth Jeevan Mary Knox May 12, 2006.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
Darryl MorrellStochastic Modeling Seminar1 Particle Filtering.
11/14  Continuation of Time & Change in Probabilistic Reasoning Project 4 progress? Grade Anxiety? Make-up Class  On Monday?  On Wednesday?
Sequential Monte Carlo and Particle Filtering Frank Wood Gatsby, November 2007 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this.
Particle Filtering. Sensors and Uncertainty Real world sensors are noisy and suffer from missing data (e.g., occlusions, GPS blackouts) Use sensor models.
Learning Models of Relational Stochastic Processes Sumit Sanghai.
Bayesian Filtering for Robot Localization
Introduction to Monte Carlo Methods D.J.C. Mackay.
Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3
Importance Sampling ICS 276 Fall 2007 Rina Dechter.
QUIZ!!  T/F: The forward algorithm is really variable elimination, over time. TRUE  T/F: Particle Filtering is really sampling, over time. TRUE  T/F:
Object Tracking using Particle Filter
From Bayesian Filtering to Particle Filters Dieter Fox University of Washington Joint work with W. Burgard, F. Dellaert, C. Kwok, S. Thrun.
SIS Sequential Importance Sampling Advanced Methods In Simulation Winter 2009 Presented by: Chen Bukay, Ella Pemov, Amit Dvash.
Machine Learning Lecture 23: Statistical Estimation with Sampling Iain Murray’s MLSS lecture on videolectures.net:
Recap: Reasoning Over Time  Stationary Markov models  Hidden Markov models X2X2 X1X1 X3X3 X4X4 rainsun X5X5 X2X2 E1E1 X1X1 X3X3 X4X4 E2E2 E3E3.
Computer Science, Software Engineering & Robotics Workshop, FGCU, April 27-28, 2012 Fault Prediction with Particle Filters by David Hatfield mentors: Dr.
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
Summary of This Course Huanhuan Chen. Outline  Basics about Signal & Systems  Bayesian inference  PCVM  Hidden Markov Model  Kalman filter  Extended.
UIUC CS 498: Section EA Lecture #21 Reasoning in Artificial Intelligence Professor: Eyal Amir Fall Semester 2011 (Some slides from Kevin Murphy (UBC))
-Arnaud Doucet, Nando de Freitas et al, UAI
Maximum a posteriori sequence estimation using Monte Carlo particle filters S. J. Godsill, A. Doucet, and M. West Annals of the Institute of Statistical.
QUIZ!!  In HMMs...  T/F:... the emissions are hidden. FALSE  T/F:... observations are independent given no evidence. FALSE  T/F:... each variable X.
Tractable Inference for Complex Stochastic Processes X. Boyen & D. Koller Presented by Shiau Hong Lim Partially based on slides by Boyen & Koller at UAI.
Sequential Monte-Carlo Method -Introduction, implementation and application Fan, Xin
1 Chapter 15 Probabilistic Reasoning over Time. 2 Outline Time and UncertaintyTime and Uncertainty Inference: Filtering, Prediction, SmoothingInference:
Particle Filtering. Sensors and Uncertainty Real world sensors are noisy and suffer from missing data (e.g., occlusions, GPS blackouts) Use dynamics models.
Tracking with dynamics
Probability and Time. Overview  Modelling Evolving Worlds with Dynamic Baysian Networks  Simplifying Assumptions Stationary Processes, Markov Assumption.
Particle Filtering. Sensors and Uncertainty Real world sensors are noisy and suffer from missing data (e.g., occlusions, GPS blackouts) Use sensor models.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Rao-Blackwellised Particle Filtering for Dynamic Bayesian Network Arnaud Doucet Nando de Freitas Kevin Murphy Stuart Russell.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Daphne Koller Overview Conditional Probability Queries Probabilistic Graphical Models Inference.
Rao-Blackwellised Particle Filtering for Dynamic Bayesian Networks Arnaud Doucet, Nando de Freitas, Kevin Murphy and Stuart Russell CS497EA presentation.
Dynamic Bayesian Network Fuzzy Systems Lifelog management.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Instructor: Eyal Amir Grad TAs: Wen Pu, Yonatan Bisk Undergrad TAs: Sam Johnson, Nikhil Johri CS 440 / ECE 448 Introduction to Artificial Intelligence.
CS 541: Artificial Intelligence Lecture VIII: Temporal Probability Models.
CS 541: Artificial Intelligence Lecture VIII: Temporal Probability Models.
CS498-EA Reasoning in AI Lecture #23 Instructor: Eyal Amir Fall Semester 2011.
Particle Filters: Theory and applications to mobile robot localization and map building Jose Luis Blanco Claraco August 2005 System Engineering and Automation.
Introduction to Sampling based inference and MCMC
HMM: Particle filters Lirong Xia. HMM: Particle filters Lirong Xia.
Course: Autonomous Machine Learning
Introduction to particle filter
Probabilistic Reasoning over Time
Visual Tracking CMPUT 615 Nilanjan Ray.
Auxiliary particle filtering: recent developments
Particle Filtering ICS 275b 2002.
Introduction to particle filter
Particle Filtering.
HMM: Particle filters Lirong Xia. HMM: Particle filters Lirong Xia.
Presentation transcript:

Dynamic Bayesian Networks and Particle Filtering COMPSCI 276 (chapter 15, Russel and Norvig) 2007

Dynamic Belief Networks (DBNs) Bayesian Network at time t Bayesian Network at time t+1 Transition arcs XtXt X t+1 YtYt Y t+1 X0X0 X1X1 X2X2 Y0Y0 Y1Y1 Y2Y2 Unrolled DBN for t=0 to t=10 X 10 Y 10

Dynamic Belief Networks (DBNs) Two-stage influence diagram Interaction graph

Notation X t – value of X at time t X 0:t ={X 0,X 1,…,X t }– vector of values of X Y t – evidence at time t Y 0:t = {Y 0,Y 1,…,Y t } X0X0 X1X1 X2X2 Y0Y0 Y1Y1 Y2Y2 DBN t=0 t=1t=2 XtXt X t+1 YtYt Y t+1 t=1t=2 2-time slice

Inference is hard, need approximation Mini-bucket? Sampling?

Same Queries Compute P(X 0:t |Y 0:t ) or P(X t |Y 0:t ) –Example P(X 0:10 |Y 0:10 ) or P(X 10 |Y 0:10 ) –Filtering –Prediction –Smoothing –MPE Hard!!! over a long time period Approximate! Sample!

Particle Filtering (PF) = “ condensation ” = “ sequential Monte Carlo ” = “ survival of the fittest ” –PF can treat any type of probability distribution, non-linearity, and non-stationarity; –PF are powerful sampling based inference/learning algorithms for DBNs.

Particle Filtering

Example Particle t ={a t,b t,c t }

PF Sampling Particle (t) ={a t,b t,c t } Compute particle (t+1): Sample b t+1, from P(b|a t,c t ) Sample a t+1, from P(a|b t+1,c t ) Sample c t+1, from P(c|b t+1,a t+1 ) Weight particle w t+1 If weight is too small, discard Otherwise, multiply

Drawback of PF –Inefficient in high-dimensional spaces (Variance becomes so large) Solution –Rao-Balckwellisation, that is, sample a subset of the variables allowing the remainder to be integrated out exactly. The resulting estimates can be shown to have lower variance. Rao-Blackwell Theorem Drawback of PF

Problem Formulation Model : general state space model/DBN with hidden variables z t and observed variables y t Objective: –or filtering density –To solve this problem,one need approximation schemes because of intractable integrals

Assume conditional posterior distribution p(x 0:t | y 1:t,r 0:t,) is analytically tractable We only need to focus on estimating p(r 0:t | y 1:t ), which lies in a space of reduced dimension: Rao-Blackwellised PF Divide hidden variables into two groups: r t and x t

Example Sample Only B t

Importance Sampling and Rao- Blackwellisation Monte Carlo integration