Rao-Blackwellised Particle Filtering Based on Rao-Blackwellised Particle Filtering for Dynamic Bayesian Networks by Arnaud Doucet, Nando de Freitas, Kevin.

Slides:



Advertisements
Similar presentations
State Estimation and Kalman Filtering CS B659 Spring 2013 Kris Hauser.
Advertisements

Rutgers CS440, Fall 2003 Review session. Rutgers CS440, Fall 2003 Topics Final will cover the following topics (after midterm): 1.Uncertainty & introduction.
Dynamic Bayesian Networks (DBNs)
Lirong Xia Hidden Markov Models Tue, March 28, 2014.
Lirong Xia Approximate inference: Particle filter Tue, April 1, 2014.
Hidden Markov Models Reading: Russell and Norvig, Chapter 15, Sections
Advanced Artificial Intelligence
1 Reasoning Under Uncertainty Over Time CS 486/686: Introduction to Artificial Intelligence Fall 2013.
10/28 Temporal Probabilistic Models. Temporal (Sequential) Process A temporal process is the evolution of system state over time Often the system state.
CS 188: Artificial Intelligence Fall 2009 Lecture 20: Particle Filtering 11/5/2009 Dan Klein – UC Berkeley TexPoint fonts used in EMF. Read the TexPoint.
Graphical Models for Mobile Robot Localization Shuang Wu.
Part 3 of 3: Beliefs in Probabilistic Robotics. References and Sources of Figures Part 1: Stuart Russell and Peter Norvig, Artificial Intelligence, 2.
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
Adaptive Rao-Blackwellized Particle Filter and It’s Evaluation for Tracking in Surveillance Xinyu Xu and Baoxin Li, Senior Member, IEEE.
SLAM: Simultaneous Localization and Mapping: Part I Chang Young Kim These slides are based on: Probabilistic Robotics, S. Thrun, W. Burgard, D. Fox, MIT.
Nonlinear and Non-Gaussian Estimation with A Focus on Particle Filters Prasanth Jeevan Mary Knox May 12, 2006.
Bayesian Networks. Graphical Models Bayesian networks Conditional random fields etc.
Part 2 of 3: Bayesian Network and Dynamic Bayesian Network.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
11/14  Continuation of Time & Change in Probabilistic Reasoning Project 4 progress? Grade Anxiety? Make-up Class  On Monday?  On Wednesday?
Announcements Homework 8 is out Final Contest (Optional)
CS 188: Artificial Intelligence Fall 2009 Lecture 19: Hidden Markov Models 11/3/2009 Dan Klein – UC Berkeley.
Dynamic Bayesian Networks CSE 473. © Daniel S. Weld Topics Agency Problem Spaces Search Knowledge Representation Reinforcement Learning InferencePlanningLearning.
Bayesian Filtering for Robot Localization
Markov Localization & Bayes Filtering
QUIZ!!  T/F: The forward algorithm is really variable elimination, over time. TRUE  T/F: Particle Filtering is really sampling, over time. TRUE  T/F:
Object Tracking using Particle Filter
From Bayesian Filtering to Particle Filters Dieter Fox University of Washington Joint work with W. Burgard, F. Dellaert, C. Kwok, S. Thrun.
Dynamic Probabilistic Relational Models Paper by: Sumit Sanghai, Pedro Domingos, Daniel Weld Anna Yershova Presentation slides are adapted.
Recap: Reasoning Over Time  Stationary Markov models  Hidden Markov models X2X2 X1X1 X3X3 X4X4 rainsun X5X5 X2X2 E1E1 X1X1 X3X3 X4X4 E2E2 E3E3.
1 Robot Environment Interaction Environment perception provides information about the environment’s state, and it tends to increase the robot’s knowledge.
UIUC CS 498: Section EA Lecture #21 Reasoning in Artificial Intelligence Professor: Eyal Amir Fall Semester 2011 (Some slides from Kevin Murphy (UBC))
Dynamic Bayesian Networks and Particle Filtering COMPSCI 276 (chapter 15, Russel and Norvig) 2007.
Announcements  Project 3 will be posted ASAP: hunting ghosts with sonar.
-Arnaud Doucet, Nando de Freitas et al, UAI
Bayes’ Nets: Sampling [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available.
Kalman Filter Notes Prateek Tandon.
4. Particle Filtering For DBLOG PF, regular BLOG inference in each particle Open-Universe State Estimation with DBLOG Rodrigo de Salvo Braz*, Erik Sudderth,
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
QUIZ!!  In HMMs...  T/F:... the emissions are hidden. FALSE  T/F:... observations are independent given no evidence. FALSE  T/F:... each variable X.
Tractable Inference for Complex Stochastic Processes X. Boyen & D. Koller Presented by Shiau Hong Lim Partially based on slides by Boyen & Koller at UAI.
Bayesian networks and their application in circuit reliability estimation Erin Taylor.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
1 Bayesian Networks: A Tutorial. 2 Introduction Suppose you are trying to determine if a patient has tuberculosis. You observe the following symptoms:
CPSC 422, Lecture 17Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 17 Oct, 19, 2015 Slide Sources D. Koller, Stanford CS - Probabilistic.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Rao-Blackwellised Particle Filtering for Dynamic Bayesian Network Arnaud Doucet Nando de Freitas Kevin Murphy Stuart Russell.
04/21/2005 CS673 1 Being Bayesian About Network Structure A Bayesian Approach to Structure Discovery in Bayesian Networks Nir Friedman and Daphne Koller.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
Rao-Blackwellised Particle Filtering for Dynamic Bayesian Networks Arnaud Doucet, Nando de Freitas, Kevin Murphy and Stuart Russell CS497EA presentation.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Weng-Keen Wong, Oregon State University © Bayesian Networks: A Tutorial Weng-Keen Wong School of Electrical Engineering and Computer Science Oregon.
CS 541: Artificial Intelligence Lecture VIII: Temporal Probability Models.
HMM: Particle filters Lirong Xia. HMM: Particle filters Lirong Xia.
Today.
Bayesian Networks: A Tutorial
Probabilistic Robotics
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Artificial Intelligence
CS 4/527: Artificial Intelligence
Introduction to particle filter
CAP 5636 – Advanced Artificial Intelligence
Probabilistic Reasoning over Time
Introduction to particle filter
CS 188: Artificial Intelligence
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Chapter14-cont..
Hidden Markov Models Lirong Xia.
HMM: Particle filters Lirong Xia. HMM: Particle filters Lirong Xia.
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Presentation transcript:

Rao-Blackwellised Particle Filtering Based on Rao-Blackwellised Particle Filtering for Dynamic Bayesian Networks by Arnaud Doucet, Nando de Freitas, Kevin Murphy, and Stuart Russel Other sources Artificial Intelligence: A Modern Approach by Stuart Russel and Peter Norvig Presented by Boris Lipchin

Introduction PF applications (localization, SLAM, etc)‏ Draw-backs/Benefits Bayes Nets Dynamic Bayes Nets Particle Filtering Rao-Blackwell PF

Bayes Net Example Adapted from Artificial Intelligence: A Modern Approach (Norvig and Russell)‏ What is the probability of Burgalry given that John calls but mary doesn't call?

Bayesian Network Digraph where edges represent conditional probabilities If A is the parent of B, B is said to be conditioned on A More compact representation than writing down full joint distribution table People rarely know absolute probability, but can predict conditional probabilities with great accuracy (i.e. doctors and symptoms)‏

Bayes Net Example Adapted from Artificial Intelligence: A Modern Approach (Norvig and Russell)‏ What is the probability of Burgalry given that John calls but mary doesn't call?

Dynamic Bayesian Networks Represent progress of a system over time 1 st Order Markov DBN: state variables can only depend on current and previous state DBNs represent temporal probability distributions Kalman Filter is a special case of a DBN Can model non-linearities (Kalman produces single multivariate Guassian)‏ Untractable to analyze

Basic DBN Example Adapted from Artificial Intelligence: A Modern Approach (Norvig and Russell)‏ Rain 0 Rain 1 Umbrella 1 Rain 0 Rain 1 Umbrella 1 Rain 2 Umbrella 2

DBN Analysis Unrolling makes DBNs just like Bayesian Network Online filtering algorithm: variable elimination As state grows, complexity of analysis per slice becomes exponential: O( d n+1 )‏ Necessity for approximate inference Particle Filtering

Constant sample count per slice achieves constant processing time Samples represent state distribution But evidence variable (umbrella) never conditions future state (rain)! Weigh future population by evidence likelihood Applications: localization, SLAM

Particle Filtering: Basic Algorithm Create initial population: Based on P( X 0 )‏ Update phase: propogate samples Transition model: P( x t+1 | x t )‏ Weigh distribution with evidence likelihood W( x t+1 | e 1:t+1 ) = P( e t+1 | x t+1 ) * N( x t+1 | e 1:t )‏ Resample to re-create unweighted population of N samples based on created weighted distribution

Visual example Adapted from Artificial Intelligence: A Modern Approach (Norvig and Russell)‏ oooo oooo ooo oo True False Rain t Rain t+1 Propogate ooo oo Rain t+1 oooo oooo Rain t+1 WeightResample ⌐umbrella This method converges assymptotically to the real distribution as N → ∞

RBPF Key concept: decrease number of particles neccessary to achieve same accuracy with regular PF Requirement: Partition state nodes Z(t) into R(t) and X(t) s.t.: P( R 1:t | Y 1:t ) can be predicted with PF P( X t | R 1:t,Y 1:t ) can be updated analytically/filtered Paper does not describe partitioning methods, efficient partitioning algorithms are assumed

RBPF: Concept Proof PF approximates P( Z 1:t | Y 1:t ) = P( R 1:t, X 1:t | Y 1:t ) Remember state space Z partitioned into R and X P( R 1:t, X 1:t | Y 1:t ) = P( R 1:t | Y 1:t ) * P( X t | R 1:t,Y 1:t )‏ By chain rule property of probability Sampling just R requires fewer particles, decreasing complexity Sampling X becomes amortized constant time

RBPF DBNs – remember arrows Adapted from Rao-Blackwellised Particle Filtering for Dynamic Bayesian Networks (Murphy and Russell 2001)‏

RBPF DBNs R(t) is called a root, and X(t) a leaf of the DBN (a) Is a canonical DBN to which RBPF can be applied (b) R(t) is a more common partitioning as it simplifies the Particle Filtering of the root in the RBPF (c) Is a convenient partitioning when some root nodes model discontinuous state changes, and others some are the parent of the observation, and model observed outliers

RBPF: Algorithm Root marginal distribution: δ is the Dirac delta function w is the weight of the i-th particle at slice t and is computed by and then normalized. Leaf marginal:

RBPF: Update, Propogate, Weigh The root particles in RBPF are propogated much like PF particles The leaf marginal propogation is computed with an optimal filter (Rao-Blackwellisation step)‏ The leaf and root nodes together compose the entire state of the system, and thus can be weighted and resampled for the next slice.

Example: Localization SLAM: P( x, m | z, u )= p( m | x, z, u)p( x | z, u ) m is the leaf, x is the root in the RBPF Particle updates based on input, expensive, we keep number of particles down posemapobservationsodometry Mapping conditioned on position and world Particle filter for position hypothesis

The Scenario Assume a world of two blue and yellow states labeled a and b (left to right)‏ A robot can successfuly move between adjacent states with a probabilty P m =.5 (transition model)‏ The robot is equipped with a color sensor that correctly identifies color with probability P c =.6

RBPF SLAM Using N = 5 particles P(X) represents state distribution (localization)‏ P(M) represents color distribution (mapping)‏ Prior for colors is an even distribution (unmapped)‏ For simplicity, P(X=a) = 1, P(X=b) = 0

RBPF SLAM Randomly select particles according to prior distribution (labeled by number)‏ arrow represents real robot position/detected color Create particles based on color 1,5,42,1,3.5.5 Remember: This means particle 1 hallucinates yellow in both boxes, particle 2 hallucinatees yellow only in right box and blue in left, so on and so forth. Represents mapping based on particle count

RBPF SLAM P(X(t)=a) = 1 Calculate weights: W1 -> P(E(a)=y | M(a)=y,M(b)=y)= P(E(a)=y)*P(M(a)=y | E(a)=y)*P(M(b)=y | E(a)=y,M(a)=y)/ ( P(M(a)=y)*P(M(b)=y | M(a)=y) ) =.5 *.6 *.5 / (.5 *.5) =.6 W2 -> P( E(a)=y | M(a)=b,M(b)=y)= P(E(a)=y)*P(M(a)=b | E(a)=y)*P(M(b)=y | E(a)=y, M(a)=b)/ ( P(M(a)=b)*P(M(b)=y | M(a)=b) )=.5 *.4 *.5 / (.5 *.5) =.4 You can calculate these guys ad nauseum 1,5,42,1,3 robot.5.5

RBPF SLAM P(X(t)=a) = 1 Next step is to resample based on weights (shown below)‏ Find P(X(t)) distribution given previous state and current map 1,5,2,41,3 robot.8.4

RBPF SLAM Calculate new weights Weigh samples and resample to obtain updated distribution for the particles estimate X(t) using optimal filter, evidence, and previous location 1,5,2,41,3 robot.8.4

RBPF SLAM: Key Ideas Imagine if X(t) was part of state space Calculations increase with number of states Number of particles RBPF simplifies calculations by giving one a ”free” localization with an optimal filter

Questions?