Localization in indoor environments by querying omnidirectional visual maps using perspective images Miguel Lourenco, V. Pedro and João P. Barreto ICRA.

Slides:



Advertisements
Similar presentations
Image Retrieval with Geometry-Preserving Visual Phrases
Advertisements

Group Meeting Presented by Wyman 10/14/2006
Aggregating local image descriptors into compact codes
Registration for Robotics Kurt Konolige Willow Garage Stanford University Patrick Mihelich JD Chen James Bowman Helen Oleynikova Freiburg TORO group: Giorgio.
Presented by Xinyu Chang
Outline SIFT Background SIFT Extraction Application in Content Based Image Search Conclusion.
TP14 - Local features: detection and description Computer Vision, FCUP, 2014 Miguel Coimbra Slides by Prof. Kristen Grauman.
Object Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition l Panoramas,
Automatic scene inference for 3D object compositing Kevin Karsch (UIUC), Sunkavalli, K. Hadap, S.; Carr, N.; Jin, H.; Fonte, R.; Sittig, M., David Forsyth.
Learning Convolutional Feature Hierarchies for Visual Recognition
IBBT – Ugent – Telin – IPI Dimitri Van Cauwelaert A study of the 2D - SIFT algorithm Dimitri Van Cauwelaert.
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Bag of Features Approach: recent work, using geometric information.
Robust and large-scale alignment Image from
Object retrieval with large vocabularies and fast spatial matching
A Study of Approaches for Object Recognition
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Feature extraction: Corners and blobs
Distinctive image features from scale-invariant keypoints. David G. Lowe, Int. Journal of Computer Vision, 60, 2 (2004), pp Presented by: Shalomi.
Object Recognition Using Distinctive Image Feature From Scale-Invariant Key point D. Lowe, IJCV 2004 Presenting – Anat Kaspi.
Scale Invariant Feature Transform (SIFT)
Blob detection.
1 Invariant Local Feature for Object Recognition Presented by Wyman 2/05/2006.
Robust estimation Problem: we want to determine the displacement (u,v) between pairs of images. We are given 100 points with a correlation score computed.
Miguel Lourenço, João P. Barreto, Abed Malti Institute for Systems and Robotics, Faculty of Science and Technology University of Coimbra, Portugal Feature.
Lecture 6: Feature matching and alignment CS4670: Computer Vision Noah Snavely.
1 Interest Operators Find “interesting” pieces of the image Multiple possible uses –image matching stereo pairs tracking in videos creating panoramas –object.
Overview Introduction to local features
Final Exam Review CS485/685 Computer Vision Prof. Bebis.
Gwangju Institute of Science and Technology Intelligent Design and Graphics Laboratory Multi-scale tensor voting for feature extraction from unstructured.
Recognition and Matching based on local invariant features Cordelia Schmid INRIA, Grenoble David Lowe Univ. of British Columbia.
Bag of Visual Words for Image Representation & Visual Search Jianping Fan Dept of Computer Science UNC-Charlotte.
Object Tracking/Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition.
Overview Harris interest points Comparing interest points (SSD, ZNCC, SIFT) Scale & affine invariant interest points Evaluation and comparison of different.
Local invariant features Cordelia Schmid INRIA, Grenoble.
Reporter: Fei-Fei Chen. Wide-baseline matching Object recognition Texture recognition Scene classification Robot wandering Motion tracking.
Lecture 4: Feature matching CS4670 / 5670: Computer Vision Noah Snavely.
Video Google: A Text Retrieval Approach to Object Matching in Videos Josef Sivic and Andrew Zisserman.
Evaluation of interest points and descriptors. Introduction Quantitative evaluation of interest point detectors –points / regions at the same relative.
Vision-based human motion analysis: An overview Computer Vision and Image Understanding(2007)
Lecture 7: Features Part 2 CS4670/5670: Computer Vision Noah Snavely.
Local invariant features Cordelia Schmid INRIA, Grenoble.
Object Recognition Based on Shape Similarity Longin Jan Latecki Computer and Information Sciences Dept. Temple Univ.,
Overview Introduction to local features Harris interest points + SSD, ZNCC, SIFT Scale & affine invariant interest point detectors Evaluation and comparison.
CS848 Similarity Search in Multimedia Databases Dr. Gisli Hjaltason Content-based Retrieval Using Local Descriptors: Problems and Issues from Databases.
Feature extraction: Corners and blobs. Why extract features? Motivation: panorama stitching We have two images – how do we combine them?
CSE 185 Introduction to Computer Vision Feature Matching.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography.
Line Matching Jonghee Park GIST CV-Lab..  Lines –Fundamental feature in many computer vision fields 3D reconstruction, SLAM, motion estimation –Useful.
776 Computer Vision Jan-Michael Frahm Spring 2012.
Recognizing specific objects Matching with SIFT Original suggestion Lowe, 1999,2004.
CSCI 631 – Foundations of Computer Vision March 15, 2016 Ashwini Imran Image Stitching.
SIFT.
776 Computer Vision Jan-Michael Frahm Spring 2012.
SIFT Scale-Invariant Feature Transform David Lowe
Lecture 07 13/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Distinctive Image Features from Scale-Invariant Keypoints
Scale Invariant Feature Transform (SIFT)
Project 1: hybrid images
TP12 - Local features: detection and description
Learning Mid-Level Features For Recognition
Cheng-Ming Huang, Wen-Hung Liao Department of Computer Science
Aim of the project Take your image Submit it to the search engine
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
SIFT.
Lecture VI: Corner and Blob Detection
Feature descriptors and matching
Presented by Xu Miao April 20, 2005
Recognition and Matching based on local invariant features
Review and Importance CS 111.
Presentation transcript:

Localization in indoor environments by querying omnidirectional visual maps using perspective images Miguel Lourenco, V. Pedro and João P. Barreto ICRA 2012

Standard Image-based Indoor Localization (1/3) How can a robot equipped with standard camera perform indoor localization? Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 2 Establishing correspondences between the query image and a database of geo-referenced images [Cummins08, Chen11, Hansen01] Query image

Standard Image-based Indoor Localization (2/3) Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 3 Querying Quantization + td-idf weighting SIFT descriptors Image Database + Inverted file Building a detailed database of environments is troublesome Number of Images StorageTime

Problem Statement Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 4 A complete coverage of the environment can be performed with an para-catadioptric camera Distortion increases the appearance difference between the images Our Contribution: A new model-based SIFT method for matching between hybrid imaging systems Query image

Presentation Outline Matching in Hybrid Imaging Systems  Comparison / drawbacks of the standard approaches  Improvements to the SIFT detector and descriptor Image-based IL using Hybrid Imaging Systems  Comparison of several database description schemes  Comparison of two searching approaches : BoV vs GVP Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 5

Matching in Hybrid Imaging Systems Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 6 Using SIFT [Lowe04] on both para- catadioptric and perspective images provides poor matching results [Puig08] 10% inliers A straightforward solution is to apply SIFT in a virtual camera perspective (VCP) [Schönbein11] Cylinder Polar Standard approaches either render a Polar [ Puig08 ] or a Cylindrical panorama [ Krishnan08 ]

Implicit cylindrical rectification - cylSIFT Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 7 Based on our previous work [Lourenco12] we propose to perform the cylindrical rectification implicitly inside SIFT framework - cylSIFT How does SIFT work ?  Image salient points detected in a scale space framework  SIFT descriptor is computed based on local image gradient Render synthetic views require to reconstruct the image signal  Interpolation artifacts severely affect SIFT performance [ Lourenco12 ]

Implicit cylindrical rectification – cylSIFT detector Render the cylinder before applying SIFT adds extra computational time and interpolation artifacts Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 8 Rectification > 2sec (Matlab ) We avoid the reconstruction artifacts by using an adaptive Gaussian filter [Lourenco12] *

Standard vs Adaptive Gaussian smoothing Inherent properties of the standard Gaussian filter  Space invariant filtering  Decouple convolution in X and Y directions Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 9 Advantages of the Simplified Adaptive Filter  Isotropic filter that can be decoupled for each image radius  A filter bank can be computed offline and loaded into memory Simplification of the adaptive filter

Implicit cylindrical rectification – cylSIFT description Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 10 Non-linear distortion modifies the local structures in the image and, by consequence, the gradients are affected Changes in local gradients of the image deteriorates SIFT descriptor performance Proposed Solution: Compute gradients in the omnidirectional image and implicitly correct them using the Jacobian matrix of the cylindrical mapping function

Detection and Matching evaluation cylSIFT completely avoids interpolation of the image signal  Better repeatability and matching with less computational burden Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 11 cylSIFT has similar performance to the VCP approach  The VCP requires a priori knowledge of the view to render to minimize viewpoints changes between the query and VCP image Set ASet B Set C Set D

Standard Image-based Indoor Localization (3/3) How can a robot use a standard camera for performing indoor localization? Compare of two image searching schemes  Standard Bags of Visual Words (BoV)  Geometry preserving Visual Phrases (GVP) [Zhang11] Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 12 Query image

BoV: Standard Bags of Visual Words (BoV) Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 13 … Length: Dictionary Size Images are represented as the histogram of words Drawback: Discard the spatial relation between words  Spatial layout can be relevant for disambiguate situations of perceptual aliasing

GVP: Encoding weak geometric constraints [Zhang11] Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 14 A B A B Offset space A B Group of Visual Words in a certain layout form a Visual Phrase GVP incorporate geometric constraints at the searching step I I’

Indoor Location Recognition - Experimental Setup Our database covers 2 teaching buildings of our campus  118 para-catadioptric images  451 perspective images are used to query the database The environment suffers from high perceptual aliasing Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 15 ? Query image

Retrieval Results Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 16 cylSIFT takes full advantage of its matching capabilities  Interpolation avoidance assure more distinctive descriptors 10% improvement of the localization success when compared with a state-of-the-art approach (Cylinder + GVP) [Chen11] Indoor recognition systems benefit with the usage of GVP  Robustness against perceptual aliasing (spatial layout matters) BoV – TOP 1GVP– TOP 1 Re-ranking - Top 5

Take home messages Interpolation artifacts affect image retrieval  cylSIFT offers better retrieval performance in hybrid imaging systems at a marginal computational cost when compared to the standard SIFT algorithm cylSIFT can be useful for other applications than localization  Hybrid fundamental matrix estimation [Puig08] First work that uses a hybrid imaging systems for image retrieval without the need of rectification Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 17

Thanks for coming Questions? Code and dataset releases: Miguel Lourenço– Institute for Systems and Robotics, Faculty of Science and Technology, University of Coimbra - 16 May Slide 18