Vision-based Navigation and Reinforcement Learning Path Finding for Social Robots Xavier Pérez *, Cecilio Angulo *, Sergio Escalera + and Diego Pardo *

Slides:



Advertisements
Similar presentations
Viktor Zhumatiya, Faustino Gomeza,
Advertisements

Application a hybrid controller to a mobile robot J.-S Chiou, K. -Y. Wang,Simulation Modelling Pratice and Theory Vol. 16 pp (2008) Professor:
Combining Detectors for Human Hand Detection Antonio Hernández, Petia Radeva and Sergio Escalera Computer Vision Center, Universitat Autònoma de Barcelona,
Visual Servo Control Tutorial Part 1: Basic Approaches Chayatat Ratanasawanya December 2, 2009 Ref: Article by Francois Chaumette & Seth Hutchinson.
Comparing Effectiveness of Bioinspired Approaches to Search and Rescue Scenarios Emily Shaeffer and Shena Cao 4/28/2011Shaeffer and Cao- ESE 313.
Víctor Ponce Miguel Reyes Xavier Baró Mario Gorga Sergio Escalera Two-level GMM Clustering of Human Poses for Automatic Human Behavior Analysis Departament.
Patch to the Future: Unsupervised Visual Prediction
Neurocomputing,Neurocomputing, Haojie Li Jinhui Tang Yi Wang Bin Liu School of Software, Dalian University of Technology School of Computer Science,
Instructor: Mircea Nicolescu Lecture 15 CS 485 / 685 Computer Vision.
Real-Time Activity Monitoring of Inpatients Miguel Reyes, Jordi Vitrià, Petia Radeva and Sergio Escalera Computer Vision Center, Universitat Autònoma de.
Nonholonomic Multibody Mobile Robots: Controllability and Motion Planning in the Presence of Obstacles (1991) Jerome Barraquand Jean-Claude Latombe.
ECE 7340: Building Intelligent Robots QUALITATIVE NAVIGATION FOR MOBILE ROBOTS Tod S. Levitt Daryl T. Lawton Presented by: Aniket Samant.
3-D Depth Reconstruction from a Single Still Image 何開暘
Probabilistic Methods in Mobile Robotics. Stereo cameras Infra-red Sonar Laser range-finder Sonar Tactiles.
Reinforcement Learning Rafy Michaeli Assaf Naor Supervisor: Yaakov Engel Visit project’s home page at: FOR.
Lecture 4: Feature matching
Color Recognition and Image Processing CSE321 MyRo Project Rene van Ee and Ankit Agarwal.
Integrating POMDP and RL for a Two Layer Simulated Robot Architecture Presented by Alp Sardağ.
GPGPU Volume Classification using SimpleOpenCL Oscar Amoros, Sergio Escalera, Anna Puig and Maria Salamó Computer Vision Center, Universitat Autònoma de.
Boğaziçi University Artificial Intelligence Lab. Artificial Intelligence Laboratory Department of Computer Engineering Boğaziçi University Techniques for.
CSE473/573 – Stereo Correspondence
Lecture 6: Feature matching and alignment CS4670: Computer Vision Noah Snavely.
Patent Liability Analysis Andrew Loveless. Potential Patent Infringement Autonomous obstacle avoidance 7,587,260 – Autonomous navigation system and method.
June 12, 2001 Jeong-Su Han An Autonomous Vehicle for People with Motor Disabilities by G. Bourhis, O.Horn, O.Habert and A. Pruski Paper Review.
Biologically Inspired Turn Control for Autonomous Mobile Robots Xavier Perez-Sala, Cecilio Angulo, Sergio Escalera.
November 25, 2014Computer Vision Lecture 20: Object Recognition IV 1 Creating Data Representations The problem with some data representations is that the.
Distinctive Image Features from Scale-Invariant Keypoints By David G. Lowe, University of British Columbia Presented by: Tim Havinga, Joël van Neerbos.
Studying Visual Attention with the Visual Search Paradigm Marc Pomplun Department of Computer Science University of Massachusetts at Boston
Constraints-based Motion Planning for an Automatic, Flexible Laser Scanning Robotized Platform Th. Borangiu, A. Dogar, A. Dumitrache University Politehnica.
Class material vs. Lab material – Lab 2, 3 vs. 4,5, 6 BeagleBoard / TI / Digilent GoPro.
DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; March Adaptive Intelligent Mobile Robotics Leslie Pack Kaelbling Artificial Intelligence.
Parallelism and Robotics: The Perfect Marriage By R.Theron,F.J.Blanco,B.Curto,V.Moreno and F.J.Garcia University of Salamanca,Spain Rejitha Anand CMPS.
Reinforcement Learning
Miguel Reyes 1,2, Gabriel Dominguez 2, Sergio Escalera 1,2 Computer Vision Center (CVC) 1, University of Barcelona (UB) 2
Using Soar for an indoor robotic search mission Scott Hanford Penn State University Applied Research Lab 1.
Study on Genetic Network Programming (GNP) with Learning and Evolution Hirasawa laboratory, Artificial Intelligence section Information architecture field.
OpenCL based machine learning labeling of biomedical datasets Oscar Amoros, Sergio Escalera and Anna Puig Computer Vision Center, Universitat Autònoma.
Wandering Standpoint Algorithm. Wandering Standpoint Algorithm for local path planning Description: –Local path planning algorithm. Required: –Local distance.
General ideas to communicate Show one particular Example of localization based on vertical lines. Camera Projections Example of Jacobian to find solution.
Visual SLAM Visual SLAM SPL Seminar (Fri) Young Ki Baik Computer Vision Lab.
University of Windsor School of Computer Science Topics in Artificial Intelligence Fall 2008 Sept 11, 2008.
Artificial Intelligence in Game Design Complex Steering Behaviors and Combining Behaviors.
Lecture 7: Features Part 2 CS4670/5670: Computer Vision Noah Snavely.
Value Function Approximation on Non-linear Manifolds for Robot Motor Control Masashi Sugiyama1)2) Hirotaka Hachiya1)2) Christopher Towell2) Sethu.
1 Research Question  Can a vision-based mobile robot  with limited computation and memory,  and rapidly varying camera positions,  operate autonomously.
yG7s#t=15 yG7s#t=15.
Turning Autonomous Navigation and Mapping Using Monocular Low-Resolution Grayscale Vision VIDYA MURALI AND STAN BIRCHFIELD CLEMSON UNIVERSITY ABSTRACT.
Jack Pinches INFO410 & INFO350 S INFORMATION SCIENCE Computer Vision I.
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
Using IR For Maze Navigation Kyle W. Lawton and Liz Shrecengost.
November 21, 2013Computer Vision Lecture 14: Object Recognition II 1 Statistical Pattern Recognition The formal description consists of relevant numerical.
Autonomous Robots Robot Path Planning (3) © Manfred Huber 2008.
2006/10/25 1 A Virtual Endoscopy System Author : Author : Anna Vilanova 、 Andreas K ö nig 、 Eduard Gr ö ller Source :Machine Graphics and Vision, 8(3),
Stereo Vision Local Map Alignment for Robot Environment Mapping Computer Vision Center Dept. Ciències de la Computació UAB Ricardo Toledo Morales (CVC)
Path Planning Based on Ant Colony Algorithm and Distributed Local Navigation for Multi-Robot Systems International Conference on Mechatronics and Automation.
Abstract LSPI (Least-Squares Policy Iteration) works well in value function approximation Gaussian kernel is a popular choice as a basis function but can.
Recognizing specific objects Matching with SIFT Original suggestion Lowe, 1999,2004.
1 Algoritmos Genéticos aplicados em Machine Learning Controle de um Robo (em inglês)
CSCI 631 – Foundations of Computer Vision March 15, 2016 Ashwini Imran Image Stitching Link: singhashwini.mesinghashwini.me.
Face recognition using Histograms of Oriented Gradients
Another Example: Circle Detection
COGNITIVE APPROACH TO ROBOT SPATIAL MAPPING
Dance robot as a tool for implementation of micro-programming
Bag-of-Visual-Words Based Feature Extraction
Reinforcement Learning in POMDPs Without Resets
Common Classification Tasks
"Playing Atari with deep reinforcement learning."
Dr. Unnikrishnan P.C. Professor, EEE
Creating Data Representations
Emir Zeylan Stylianos Filippou
Presentation transcript:

Vision-based Navigation and Reinforcement Learning Path Finding for Social Robots Xavier Pérez *, Cecilio Angulo *, Sergio Escalera + and Diego Pardo * * CETpD, UPC, Rambla de l’Exposició, 59, Vilanova i la Geltrú, Spain + Dept. Matemàtica Aplicada i Aàlisi, UB, Gran Via 585, Barcelona, Spain [1] Bay H. Ess A. Tuytelaars T. Van Gool L., Surf: Speeded up robust features, Computer Vision and Image Understanding (CVIU), 2008 [2] Peters, J. Vijayakumar, S. Schaal, S., Policy Gradient Methods for robotics, In International Conference on Intelligent Robots and Systems (IROS), 2006 [3] Peters, J., Machine Learning for Robotics: Learning Methods for Robot Motor Skills, VDM-Verlag, 2008 [4] Yang, J. Jiang Y.G. Hauptmann, A. Ngo C.W., Evaluating bag-of-visual- word representation in scene classification, MIR07 ACMMM, 2007 Sub-title 1. Overview Abstract An exportable and robust system for automatic Robot Navigation in unknown environments is proposed. The aim of the system is to allow a robot to automatically find a path that leads to a given goal, avoiding obstacles, only using vision and the least number of sensors. The system is composed by three main modules: the Artificial Vision, the Reinforcement Learning, and the Reactive Anti-collision module. Artificial Vision module is able to provide the information and functionalities needed by the Reinforcement Learning module. Visual Based navigation and a State definition are developed without the use of a map, only using proximity sensors and Sony AIBO camera images. In order to follow a route between two points in the environment, a map is usually needed to optimize the route and follow it. However, here a path finding approach is presented where a map of the environment is not needed, neither the use of artificial landmarks. 2. Reinforcement Learning According to the Reinforcement Learning (RL) paradigm, robot should take actions within its universe, looking for maximizing some notion of cumulative reward. RL algorithms attempt to find a policy that maps its current state to the actions the robot should take in those states. Formally, the basic RL model consists of: 5. CONCLUSIONS 3. State definition Computer Vision Center, Campus UAB, Edifici O, 08193, Cerdanyola Set of world states X:x є R n, where n = 53 = dictionary size + 3 sensors Set of actions U:U = [forward, backward, 90º left, 90º right] Set of scalar rewards :r є Rr є R A value of n=53 implies a high state space dimensionality, too large to grid the state space supposing that all states must be visited. It is necessary to look for a continuous RL algorithm that supports high state dimensionality, therefore, a Policy Gradient Reinforcement Learning method (PGRL) [3] is applied. The Natural Actor-Critic Algorithm described in [2] is chosen given that it supports a high state dimensionality. Actions ordered by Reinforcement Learning module must always be carry out in the same way. Therefore, it is necessary to avoid unexpected behaviors implementing reliable actions: controlled forward and controlled turn. 1.Feature extraction, applying SURF [1] on robot’s camera images. 2.To find correspondences between features from consecutive images, obtaining a set of motion vectors describing robot motion in 2D. 3.Is robot going forward or is it turning? 4. Actions (Navigation control) Similar states on the map should have similar state representations and very different state representation is due to distant or very different states. Restrictions: Only use robot sensors World’s map is not known Artificial landmarks are forbidden Idea: Robot camera image describes its position and orientation with a high level of reliability. State description: Vector containing proximity sensor values and result of apply “Bag of Visual Words”(BoVW) [4] on images from robot’s camera, using Speeded Up Robust Features (SURF ) [1] descriptor. Dictionary size: 50 In this work we presented a new approach for navigation control of mobile robots. Designed Vision based navigation works really well on Sony Aibo, and it probably could work even better on wheeled robots. The proposed system only uses the robot camera to achieve a controlled loop to go forward and another one to turn a desired angle. Besides, the robot uses proximity infrared sensors in order to avoid obstacles. Reliable state representation is obtained using proximity sensors and a 50 length histogram resulting from BoVW. Reinforcement Learning algorithm is able to work with high dimensionality data. The robot looks for the goal, producing behavior changes based on experience, without finding the optimal route that reaches the goal. It seems a reasonably useful approach despite of the needing of a better configuration for learning optimal parameters in order to achieve the desired results. Test: Image retrieval system using pictures from maze. Forward Vanishing Point (VP) is achieved looking for motion vector intersections. VP is the focus of the movement i.e. VP shows the direction of the movement of the robot. Control consists on maintain VP in the center of the image Turn 1.Turn the head in an specific angle, using neck encoders. 2.Start turning the body in the direction the head is pointing while robot keeps its head still. 3.Turn is completed when head and body are aligned. To maintain its head still, Sony Aibo tries to continue watching the same image, avoiding image modifications. Error signal to correct neck angles is steering angle: mean of parallel motion vectors.