Probabilistic Robotics Bayes Filter Implementations Gaussian filters.

Slides:



Advertisements
Similar presentations
EKF, UKF TexPoint fonts used in EMF.
Advertisements

Probabilistic Robotics
Probabilistic Robotics
Probabilistic Robotics SLAM. 2 Given: The robot’s controls Observations of nearby features Estimate: Map of features Path of the robot The SLAM Problem.
(Includes references to Brian Clipp
Robot Localization Using Bayesian Methods
IR Lab, 16th Oct 2007 Zeyn Saigol
Lab 2 Lab 3 Homework Labs 4-6 Final Project Late No Videos Write up
Markov Localization & Bayes Filtering 1 with Kalman Filters Discrete Filters Particle Filters Slides adapted from Thrun et al., Probabilistic Robotics.
Observers and Kalman Filters
Uncertainty Representation. Gaussian Distribution variance Standard deviation.
Introduction to Mobile Robotics Bayes Filter Implementations Gaussian filters.
Recursive Bayes Filtering Advanced AI
Probabilistic Robotics: Kalman Filters
Stanford CS223B Computer Vision, Winter 2007 Lecture 12 Tracking Motion Professors Sebastian Thrun and Jana Košecká CAs: Vaibhav Vaish and David Stavens.
Introduction to Kalman Filter and SLAM Ting-Wei Hsu 08/10/30.
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
Probabilistic Robotics: Motion Model/EKF Localization
Probabilistic Robotics
Stanford CS223B Computer Vision, Winter 2007 Lecture 12 Tracking Motion Professors Sebastian Thrun and Jana Košecká CAs: Vaibhav Vaish and David Stavens.
Estimation and the Kalman Filter David Johnson. The Mean of a Discrete Distribution “I have more legs than average”
Course AE4-T40 Lecture 5: Control Apllication
Probabilistic Robotics Bayes Filter Implementations Gaussian filters.
Kalman Filtering Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics TexPoint fonts used in EMF. Read.
Overview and Mathematics Bjoern Griesbach
Bayesian Filtering Dieter Fox Probabilistic Robotics Key idea: Explicit representation of uncertainty (using the calculus of probability theory) Perception.
ROBOT MAPPING AND EKF SLAM
Slam is a State Estimation Problem. Predicted belief corrected belief.
EKF and UKF Day EKF and RoboCup Soccer simulation of localization using EKF and 6 landmarks (with known correspondences) robot travels in a circular.
Bayesian Filtering for Robot Localization
Kalman filter and SLAM problem
Mobile Robot controlled by Kalman Filter
Markov Localization & Bayes Filtering
Localization and Mapping (3)
Computer vision: models, learning and inference Chapter 19 Temporal models.
From Bayesian Filtering to Particle Filters Dieter Fox University of Washington Joint work with W. Burgard, F. Dellaert, C. Kwok, S. Thrun.
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 6.2: Kalman Filter Jürgen Sturm Technische Universität München.
3D SLAM for Omni-directional Camera
Probabilistic Robotics Robot Localization. 2 Localization Given Map of the environment. Sequence of sensor measurements. Wanted Estimate of the robot’s.
Kalman Filter (Thu) Joon Shik Kim Computational Models of Intelligence.
Probabilistic Robotics Bayes Filter Implementations Gaussian filters.
Probabilistic Robotics Bayes Filter Implementations.
Young Ki Baik, Computer Vision Lab.
Mobile Robot Localization (ch. 7)
City College of New York 1 Dr. Jizhong Xiao Department of Electrical Engineering City College of New York Advanced Mobile Robotics.
1 Assignment, Project and Presentation Mobile Robot Localization by using Particle Filter by Chong Wang, Chong Fu, and Guanghui Luo. Tracking, Mapping.
State Estimation and Kalman Filtering
Unscented Kalman Filter 1. 2 Linearization via Unscented Transform EKF UKF.
State Estimation and Kalman Filtering Zeeshan Ali Sayyed.
Extended Kalman Filter
Nonlinear State Estimation
The Unscented Kalman Filter for Nonlinear Estimation Young Ki Baik.
Robust Localization Kalman Filter & LADAR Scans
Autonomous Mobile Robots Autonomous Systems Lab Zürich Probabilistic Map Based Localization "Position" Global Map PerceptionMotion Control Cognition Real.
Probabilistic Robotics Bayes Filter Implementations Gaussian filters.
Tracking We are given a contour G1 with coordinates G1={x1 , x2 , … , xN} at the initial frame t=1, were the image is It=1 . We are interested in tracking.
Probabilistic Robotics
PSG College of Technology
Unscented Kalman Filter
Unscented Kalman Filter
Probabilistic Robotics
Markov ó Kalman Filter Localization
A Short Introduction to the Bayes Filter and Related Models
Probabilistic Map Based Localization
Unscented Kalman Filter
Bayes and Kalman Filter
Probabilistic Robotics
Extended Kalman Filter
Extended Kalman Filter
Probabilistic Robotics Bayes Filter Implementations FastSLAM
Presentation transcript:

Probabilistic Robotics Bayes Filter Implementations Gaussian filters

Markov  Kalman Filter Localization Markov localization localization starting from any unknown position recovers from ambiguous situation. However, to update the probability of all positions within the whole state space at any time requires a discrete representation of the space (grid). The required memory and calculation power can thus become very important if a fine grid is used. Kalman filter localization tracks the robot and is inherently very precise and efficient. However, if the uncertainty of the robot becomes to large (e.g. collision with an object) the Kalman filter will fail and the position is definitively lost.

Kalman Filter Localization

4 Bayes Filter Reminder 1. Algorithm Bayes_filter( Bel(x),d ): 2. 0 3. If d is a perceptual data item z then 4. For all x do For all x do Else if d is an action data item u then 10. For all x do Return Bel’(x)

Prediction Correction Bayes Filter Reminder

Kalman Filter Bayes filter with Gaussians Developed in the late 1950's Most relevant Bayes filter variant in practice Applications range from economics, wheather forecasting, satellite navigation to robotics and many more. The Kalman filter "algorithm" is a couple of matrix multiplications! 6

Gaussians --   Univariate  Multivariate

Gaussians 1D 2D 3D Video

Properties of Gaussians Multivariate Univariate We stay in the “Gaussian world” as long as we start with Gaussians and perform only linear transformations

Introduction to Kalman Filter (1) Two measurements no dynamics Weighted least-square Finding minimum error After some calculation and rearrangements Another way to look at it – weigthed mean

11 Discrete Kalman Filter Estimates the state x of a discrete-time controlled process that is governed by the linear stochastic difference equation with a measurement Matrix (nxn) that describes how the state evolves from t to t-1 without controls or noise. Matrix (nxl) that describes how the control u t changes the state from t to t-1. Matrix (kxn) that describes how to map the state x t to an observation z t. Random variables representing the process and measurement noise that are assumed to be independent and normally distributed with covariance R t and Q t respectively.

12 Kalman Filter Updates in 1D predictionmeasurement correction It's a weighted mean!

13 Kalman Filter Updates in 1D

14 Kalman Filter Updates in 1D

15 Kalman Filter Updates

16 Linear Gaussian Systems: Initialization Initial belief is normally distributed:

17 Dynamics are linear function of state and control plus additive noise: Linear Gaussian Systems: Dynamics

18 Linear Gaussian Systems: Dynamics

19 Observations are linear function of state plus additive noise: Linear Gaussian Systems: Observations

20 Linear Gaussian Systems: Observations

21 Kalman Filter Algorithm 1. Algorithm Kalman_filter(  t-1,  t-1, u t, z t ): 2. Prediction: Correction: Return  t,  t

Kalman Filter Algorithm

 Prediction  Observation  Matching  Correction

24 The Prediction-Correction-Cycle Prediction

25 The Prediction-Correction-Cycle Correction

26 The Prediction-Correction-Cycle Correction Prediction

27 Kalman Filter Summary Highly efficient: Polynomial in measurement dimensionality k and state dimensionality n: O(k n 2 ) Optimal for linear Gaussian systems! Most robotics systems are nonlinear!

28 Nonlinear Dynamic Systems Most realistic robotic problems involve nonlinear functions To be continued

29 Linearity Assumption Revisited

30 Non-linear Function

31 EKF Linearization (1)

32 EKF Linearization (2)

33 EKF Linearization (3) Depends on uncertainty

34 Prediction: Correction: EKF Linearization: First Order Taylor Series Expansion

35 EKF Algorithm 1.Extended_Kalman_filter (  t-1,  t-1, u t, z t ): 2. Prediction: Correction: Return  t,  t

Kalman Filter for Mobile Robot Localization Robot Position Prediction In a first step, the robots position at time step k+1 is predicted based on its old location (time step k) and its movement due to the control input u(k): f: Odometry function

Kalman Filter for Mobile Robot Localization Robot Position Prediction: Example Odometry

Kalman Filter for Mobile Robot Localization Observation The second step it to obtain the observation Z(k+1) (measurements) from the robot’s sensors at the new location at time k+1 The observation usually consists of a set n 0 of single observations z j (k+1) extracted from the different sensors signals. It can represent raw data scans as well as features like lines, doors or any kind of landmarks. The parameters of the targets are usually observed in the sensor frame {S}. Therefore the observations have to be transformed to the world frame {W} or the measurement prediction have to be transformed to the sensor frame {S}. This transformation is specified in the function h i (seen later) – measurement prediction

Kalman Filter for Mobile Robot Localization Observation: Example Raw Date of Laser Scanner Extracted Lines in Model Space Sensor (robot) frame Set of discrete line features

Kalman Filter for Mobile Robot Localization Measurement Prediction In the next step we use the predicted robot position and the map M(k) to generate multiple predicted observations z t – what you should see They have to be transformed into the sensor frame We can now define the measurement prediction as the set containing all n i predicted observations The function h i is mainly the coordinate transformation between the world frame and the sensor frame Need to compute the measurement Jacobian to also predict uncertainties

Kalman Filter for Mobile Robot Localization Measurement Prediction: Example For prediction, only the walls that are in the field of view of the robot are selected. This is done by linking the individual lines to the nodes of the path

Kalman Filter for Mobile Robot Localization Measurement Prediction: Example The generated measurement predictions have to be transformed to the robot frame {R} According to the figure in previous slide the transformation is given by and its Jacobian by

Kalman Filter for Mobile Robot Localization Matching Assignment from observations z j (k+1) (gained by the sensors) to the targets z t (stored in the map) For each measurement prediction for which an corresponding observation is found we calculate the innovation: and its innovation covariance found by applying the error propagation law: The validity of the correspondence between measurement and prediction can e.g. be evaluated through the Mahalanobis distance:

Kalman Filter for Mobile Robot Localization Matching: Example

To find correspondence (pairs) of predicted and observed features we use the Mahalanobis distance with

Kalman Filter for Mobile Robot Localization Estimation: Applying the Kalman Filter Kalman filter gain: Update of robot’s position estimate: The associate variance

Kalman Filter for Mobile Robot Localization Estimation: 1D Case For the one-dimensional case with we can show that the estimation corresponds to the Kalman filter for one-dimension presented earlier.

Kalman Filter for Mobile Robot Localization Estimation: Example Kalman filter estimation of the new robot position : By fusing the prediction of robot position (magenta) with the innovation gained by the measurements (green) we get the updated estimate of the robot position (red)

Autonomous Indoor Navigation (Pygmalion EPFL) very robust on-the-fly localization one of the first systems with probabilistic sensor fusion 47 steps,78 meter length, realistic office environment, conducted 16 times > 1km overall distance partially difficult surfaces (laser), partially few vertical edges (vision)

50 Localization Given Map of the environment. Sequence of sensor measurements. Wanted Estimate of the robot’s position. Problem classes Position tracking Global localization Kidnapped robot problem (recovery) “Using sensory information to locate the robot in its environment is the most fundamental problem to providing a mobile robot with autonomous capabilities.” [Cox ’91]

51 Landmark-based Localization

52 1.EKF_localization (  t-1,  t-1, u t, z t, m): Prediction: Motion noise Jacobian of g w.r.t location Predicted mean Predicted covariance Jacobian of g w.r.t control

53 1.EKF_localization (  t-1,  t-1, u t, z t, m): Correction: Predicted measurement mean Pred. measurement covariance Kalman gain Updated mean Updated covariance Jacobian of h w.r.t location

54 EKF Prediction Step

55 EKF Observation Prediction Step

56 EKF Correction Step

57 Estimation Sequence (1)

58 Estimation Sequence (2)

59 Comparison to GroundTruth

60 EKF Summary Highly efficient: Polynomial in measurement dimensionality k and state dimensionality n: O(k n 2 ) Not optimal! Can diverge if nonlinearities are large! Works surprisingly well even when all assumptions are violated!

61 Linearization via Unscented Transform EKF UKF

62 UKF Sigma-Point Estimate (2) EKF UKF

63 UKF Sigma-Point Estimate (3) EKF UKF

64 Unscented Transform Sigma points Weights Pass sigma points through nonlinear function Recover mean and covariance

65 UKF_localization (  t-1,  t-1, u t, z t, m): Prediction: Motion noise Measurement noise Augmented state mean Augmented covariance Sigma points Prediction of sigma points Predicted mean Predicted covariance

66 UKF_localization (  t-1,  t-1, u t, z t, m): Correction: Measurement sigma points Predicted measurement mean Pred. measurement covariance Cross-covariance Kalman gain Updated mean Updated covariance

67 1.EKF_localization (  t-1,  t-1, u t, z t, m): Correction: Predicted measurement mean Pred. measurement covariance Kalman gain Updated mean Updated covariance Jacobian of h w.r.t location

68 UKF Prediction Step

69 UKF Observation Prediction Step

70 UKF Correction Step

71 EKF Correction Step

72 Estimation Sequence EKF PF UKF

73 Estimation Sequence EKF UKF

74 Prediction Quality EKF UKF

75 UKF Summary Highly efficient: Same complexity as EKF, with a constant factor slower in typical practical applications Better linearization than EKF: Accurate in first two terms of Taylor expansion (EKF only first term) Derivative-free: No Jacobians needed Still not optimal!

76 [Arras et al. 98]: Laser range-finder and vision High precision (<1cm accuracy) Kalman Filter-based System [Courtesy of Kai Arras]

Autonomous Indoor Navigation (Pygmalion EPFL) very robust on-the-fly localization one of the first systems with probabilistic sensor fusion 47 steps,78 meter length, realistic office environment, conducted 16 times > 1km overall distance partially difficult surfaces (laser), partially few vertical edges (vision)

78 Multi- hypothesis Tracking

79 Belief is represented by multiple hypotheses Each hypothesis is tracked by a Kalman filter Additional problems: Data association: Which observation corresponds to which hypothesis? Hypothesis management: When to add / delete hypotheses? Huge body of literature on target tracking, motion correspondence etc. Localization With MHT

80 Hypotheses are extracted from LRF scans Each hypothesis has probability of being the correct one: Hypothesis probability is computed using Bayes’ rule Hypotheses with low probability are deleted. New candidates are extracted from LRF scans. MHT: Implemented System (1) [Jensfelt et al. ’00]

81 MHT: Implemented System (2) Courtesy of P. Jensfelt and S. Kristensen

82 MHT: Implemented System (3) Example run Map and trajectory # hypotheses #hypotheses vs. time P(H best ) Courtesy of P. Jensfelt and S. Kristensen