Behrouz Haji Soleimani Dr. Moradi. Outline What is uncertainty? Some examples Solutions to uncertainty Ignoring uncertainty Markov Decision Process (MDP)

Slides:



Advertisements
Similar presentations
Value and Planning in MDPs. Administrivia Reading 3 assigned today Mahdevan, S., “Representation Policy Iteration”. In Proc. of 21st Conference on Uncertainty.
Advertisements

Markov Decision Processes (MDPs) read Ch utility-based agents –goals encoded in utility function U(s), or U:S  effects of actions encoded in.
1 University of Southern California Keep the Adversary Guessing: Agent Security by Policy Randomization Praveen Paruchuri University of Southern California.
Partially Observable Markov Decision Process (POMDP)
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Solving POMDPs Using Quadratically Constrained Linear Programs Christopher Amato.
ANDREW MAO, STACY WONG Regrets and Kidneys. Intro to Online Stochastic Optimization Data revealed over time Distribution of future events is known Under.
Decision Theoretic Planning
Computational Stochastic Optimization:
Creating High-quality Roadmaps for Motion Planning in Virtual Environments Roland Geraerts and Mark Overmars IROS 2006.
Probabilistic Roadmap
A Hybridized Planner for Stochastic Domains Mausam and Daniel S. Weld University of Washington, Seattle Piergiorgio Bertoli ITC-IRST, Trento.
An Introduction to Markov Decision Processes Sarah Hickmott
Planning under Uncertainty
Planning in surgery and surgical simulation Comp course presentation Mert Sedef.
© 2003 Warren B. Powell Slide 1 Approximate Dynamic Programming for High Dimensional Resource Allocation NSF Electric Power workshop November 3, 2003 Warren.
SA-1 1 Probabilistic Robotics Planning and Control: Markov Decision Processes.
1 Single Robot Motion Planning - II Liang-Jun Zhang COMP Sep 24, 2008.
Tracking a moving object with real-time obstacle avoidance Chung-Hao Chen, Chang Cheng, David Page, Andreas Koschan and Mongi Abidi Imaging, Robotics and.
INSTITUTO DE SISTEMAS E ROBÓTICA Minimax Value Iteration Applied to Robotic Soccer Gonçalo Neto Institute for Systems and Robotics Instituto Superior Técnico.
Bayesian Reinforcement Learning with Gaussian Processes Huanren Zhang Electrical and Computer Engineering Purdue University.
1 Motion Planning for Multiple Robots B. Aronov, M. de Berg, A. Frank van der Stappen, P. Svestka, J. Vleugels Presented by Tim Bretl.
Source-Destination Routing Optimal Strategies Eric Chi EE228a, Fall 2002 Dept. of EECS, U.C. Berkeley.
Planning in MDPs S&B: Sec 3.6; Ch. 4. Administrivia Reminder: Final project proposal due this Friday If you haven’t talked to me yet, you still have the.
Concurrent Probabilistic Temporal Planning (CPTP) Mausam Joint work with Daniel S. Weld University of Washington Seattle.
Discretization Pieter Abbeel UC Berkeley EECS
CS 326A: Motion Planning Kynodynamic Planning + Dealing with Moving Obstacles + Dealing with Uncertainty + Dealing with Real-Time Issues.
Planning to learn. Progress report Last time: Transition functions & stochastic outcomes Markov chains MDPs defined Today: Exercise completed Value functions.
Markov Decision Processes Value Iteration Pieter Abbeel UC Berkeley EECS TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.:
Department of Computer Science Undergraduate Events More
Stochastic Roadmap Simulation: An Efficient Representation and Algorithm for Analyzing Molecular Motion Mehmet Serkan Apaydin, Douglas L. Brutlag, Carlos.
More RL. MDPs defined A Markov decision process (MDP), M, is a model of a stochastic, dynamic, controllable, rewarding process given by: M = 〈 S, A,T,R.
Exploration in Reinforcement Learning Jeremy Wyatt Intelligent Robotics Lab School of Computer Science University of Birmingham, UK
Presentation for the course : Advanced Robotics Behdad Soleimani 9 March 2009.
Instructor: Vincent Conitzer
MAKING COMPLEX DEClSlONS
The Stochastic Motion Roadmap: A Sampling Framework for Planning with Markov Motion Uncertainty Changsi An You-Wei Cheah.
General Polynomial Time Algorithm for Near-Optimal Reinforcement Learning Duke University Machine Learning Group Discussion Leader: Kai Ni June 17, 2005.
Combined Task and Motion Planning for Mobile Manipulation Jason Wolfe, etc. University of California, Berkeley Mozi Song INFOTECH 1.
TKK | Automation Technology Laboratory Partially Observable Markov Decision Process (Chapter 15 & 16) José Luis Peralta.
Introduction to Reinforcement Learning Dr Kathryn Merrick 2008 Spring School on Optimisation, Learning and Complexity Friday 7 th.
Utilities and MDP: A Lesson in Multiagent System Based on Jose Vidal’s book Fundamentals of Multiagent Systems Henry Hexmoor SIUC.
Modeling & Planning Deniz Güven Needle Insertion.
CPSC 7373: Artificial Intelligence Lecture 10: Planning with Uncertainty Jiang Bian, Fall 2012 University of Arkansas at Little Rock.
Privacy-Preserving Bayes-Adaptive MDPs CS548 Term Project Kanghoon Lee, AIPR Lab., KAIST CS548 Advanced Information Security Spring 2010.
By: Messias, Spaan, Lima Presented by: Mike Plasker DMES – Ocean Engineering.
Department of Computer Science Undergraduate Events More
Reinforcement Learning Yishay Mansour Tel-Aviv University.
A Tutorial on the Partially Observable Markov Decision Process and Its Applications Lawrence Carin June 7,2006.
Learning to Navigate Through Crowded Environments Peter Henry 1, Christian Vollmer 2, Brian Ferris 1, Dieter Fox 1 Tuesday, May 4, University of.
Decision Theoretic Planning. Decisions Under Uncertainty  Some areas of AI (e.g., planning) focus on decision making in domains where the environment.
CS 484 – Artificial Intelligence1 Announcements Homework 5 due Tuesday, October 30 Book Review due Tuesday, October 30 Lab 3 due Thursday, November 1.
Reinforcement Learning Dynamic Programming I Subramanian Ramamoorthy School of Informatics 31 January, 2012.
Department of Computer Science Undergraduate Events More
Planning Under Uncertainty. Sensing error Partial observability Unpredictable dynamics Other agents.
Stut 11 Robot Path Planning in Unknown Environments Using Particle Swarm Optimization Leandro dos Santos Coelho and Viviana Cocco Mariani.
Prof. Dr. Holger Schlingloff 1,2 Dr. Esteban Pavese 1
CS b659: Intelligent Robotics
Analytics and OR DP- summary.
Markov Decision Processes
Markov Decision Processes
Course Logistics CS533: Intelligent Agents and Decision Making
Sampling and Connection Strategies for Probabilistic Roadmaps
Markov Decision Problems
CS 416 Artificial Intelligence
CS 416 Artificial Intelligence
Chapter 4 . Trajectory planning and Inverse kinematics
Reinforcement Nisheeth 18th January 2019.
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 3
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
Presentation transcript:

Behrouz Haji Soleimani Dr. Moradi

Outline What is uncertainty? Some examples Solutions to uncertainty Ignoring uncertainty Markov Decision Process (MDP) Stochastic Motion Roadmap A detailed example Conclusion

What is uncertainty? Uncertainty in sensing the current state of the robot and workspace is not known with certainty Predictability the future state of the robot and workspace cannot be deterministically predicted even when the current state and future actions are known

Uncertainty in sensing It is not the world that is imperfect, it is our knowledge of it

Predictability Uncertainty in workspace Uncertainty in goal location Dynamic environments with moving obstacles Uncertainty in robot’s motion

Uncertainty example A robot with imperfect sensing must reach a goal location among moving obstacles (dynamic world)

Uncertainty example Robot created at Stanford’s ARL Lab to study issues in robot control and planning in no-gravity space environment air thrusters gas tank air bearing

Uncertainty in motion

Markov Decision Process (MDP) MDP is a general approach to considering uncertainty Determines model of the environment Descretizes state space Requires explicitly defining transition probabilities between states We can use dynamic programming to solve the MDP

Stochastic Motion Roadmap Combines a roadmap representation of configuration space with the theory of MDP’s Maximizes the probability of success Uses sampling to learn the configuration space (represented as states) learn the stochastic motion model (represented as state transition probabilities) Discretizes state space Discretizes actions

Stochastic Motion Roadmap Learning Phase Selecting random sample of discrete states Sample the robot’s motion model to build a Stochastic Motion Roadmap (SMR) Calculating transition probabilities for each action Query Phase Specify initial and goal states Roadmap is used to find a feasible path Possibly optimizing some criteria such as minimum length

Building the roadmap

Maximizing probability of success build an n × n transition probability matrix P(u) for each u U For each tuple (s, t, p), we set equals the probability of transitioning from state s to state t given that action u is performed

Maximizing probability of success

It is an MDP and has the form of the Bellman equation Where and It can be optimally solved using infinite horizon dynamic programming

A detailed example

Conclusion Uncertainty has a great effect on successfully reaching the goal MDP can consider uncertainty in the model SMR combines PRM and MDP to handle uncertainty SMR maximizes the probability of success SMR makes balance between path safety and minimum length Continuous actions in SMR is still an open question

References [1] R. Alterovitz, T. Simeon, and K. Goldberg, “The Stochastic Motion Roadmap: A Sampling Framework for Planning with Markov Motion Uncertainty” 2007 [2] R. Alterovitz, M. Branicky, and K. Goldberg, “Constant-curvature motion planning under uncertainty with applications in image-guided medical needle steering,” in Workshop on the Algorithmic Foundations of Robotics, July [3] R. Alterovitz, A. Lim, K. Goldberg, G. S. Chirikjian, and A. M. Okamura, “Steering flexible needles under Markov motion uncertainty,” in Proc. IEEE/RSJ Int. Conf. on Intelligent Robots and Systems (IROS), Aug. 2005, pp. 120–125. [4] B. Bouilly, T. Simeon, and R. Alami, “A numerical technique for planning motion strategies of a mobile robot in presence of uncertainty,” in Proc. IEEE Int. Conf. on Robotics and Automation (ICRA), Nagoya, Japan, May 1995.

Questions ?

Thank you