3-D Depth Reconstruction from a Single Still Image 何開暘 2010.6.11.

Slides:



Advertisements
Similar presentations
Feature Detection. Description Localization More Points Robust to occlusion Works with less texture More Repeatable Robust detection Precise localization.
Advertisements

OpenCV Introduction Hang Xiao Oct 26, History  1999 Jan : lanched by Intel, real time machine vision library for UI, optimized code for intel 
November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
Presented By: Vennela Sunnam
嵌入式視覺 Feature Extraction
TP14 - Local features: detection and description Computer Vision, FCUP, 2014 Miguel Coimbra Slides by Prof. Kristen Grauman.
July 27, 2002 Image Processing for K.R. Precision1 Image Processing Training Lecture 1 by Suthep Madarasmi, Ph.D. Assistant Professor Department of Computer.
Computer Vision Lecture 16: Texture
1 Video Processing Lecture on the image part (8+9) Automatic Perception Volker Krüger Aalborg Media Lab Aalborg University Copenhagen
Foreground Modeling The Shape of Things that Came Nathan Jacobs Advisor: Robert Pless Computer Science Washington University in St. Louis.
Shape-from-X Class 11 Some slides from Shree Nayar. others.
1Ellen L. Walker Edges Humans easily understand “line drawings” as pictures.
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
Content Based Image Retrieval
A Novel 2D-to-3D Conversion System Using Edge Information IEEE Transactions on Consumer Electronics 2010 Chao-Chung Cheng Chung-Te li Liang-Gee Chen.
1 Visual Information Extraction in Content-based Image Retrieval System Presented by: Mian Huang Weichuan Dong Apr 29, 2004.
Introduction to Computer Vision 3D Vision Topic 9 Stereo Vision (I) CMPSCI 591A/691A CMPSCI 570/670.
Texture Reading: Chapter 9 (skip 9.4) Key issue: How do we represent texture? Topics: –Texture segmentation –Texture-based matching –Texture synthesis.
CS292 Computational Vision and Language Visual Features - Colour and Texture.
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
High Speed Obstacle Avoidance using Monocular Vision and Reinforcement Learning Jeff Michels Ashutosh Saxena Andrew Y. Ng Stanford University ICML 2005.
A Novel 2D To 3D Image Technique Based On Object- Oriented Conversion.
May 2004Stereo1 Introduction to Computer Vision CS / ECE 181B Tuesday, May 11, 2004  Multiple view geometry and stereo  Handout #6 available (check with.
3D/Multview Video. Outline Introduction 3D Perception and HVS 3D Displays 3D Video Representation Compression.
Computer vision.
A Bayesian Approach For 3D Reconstruction From a Single Image
1 Interest Operators Harris Corner Detector: the first and most basic interest operator Kadir Entropy Detector and its use in object recognition SIFT interest.
THE UNIVERSITY OF BRITISH COLUMBIA Random Forests-Based 2D-to- 3D Video Conversion Presenter: Mahsa Pourazad M. Pourazad, P. Nasiopoulos, and A. Bashashati.
Local invariant features Cordelia Schmid INRIA, Grenoble.
(Wed) Young Ki Baik Computer Vision Lab.
December 4, 2014Computer Vision Lecture 22: Depth 1 Stereo Vision Comparing the similar triangles PMC l and p l LC l, we get: Similarly, for PNC r and.
1 Perception, Illusion and VR HNRS 299, Spring 2008 Lecture 8 Seeing Depth.
Supervised Learning of Edges and Object Boundaries Piotr Dollár Zhuowen Tu Serge Belongie.
Course 9 Texture. Definition: Texture is repeating patterns of local variations in image intensity, which is too fine to be distinguished. Texture evokes.
1 Artificial Intelligence: Vision Stages of analysis Low level vision Surfaces and distance Object Matching.
Computer Vision Lecture #10 Hossam Abdelmunim 1 & Aly A. Farag 2 1 Computer & Systems Engineering Department, Ain Shams University, Cairo, Egypt 2 Electerical.
Local invariant features Cordelia Schmid INRIA, Grenoble.
Epitomic Location Recognition A generative approach for location recognition K. Ni, A. Kannan, A. Criminisi and J. Winn In proc. CVPR Anchorage,
Chapter 5 Multi-Cue 3D Model- Based Object Tracking Geoffrey Taylor Lindsay Kleeman Intelligent Robotics Research Centre (IRRC) Department of Electrical.
Edge Detection and Geometric Primitive Extraction Jinxiang Chai.
Autonomous Robots Vision © Manfred Huber 2014.
Local features: detection and description
A New Method for Crater Detection Heather Dunlop November 2, 2006.
1Ellen L. Walker 3D Vision Why? The world is 3D Not all useful information is readily available in 2D Why so hard? “Inverse problem”: one image = many.
Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth
Stereo Vision Local Map Alignment for Robot Environment Mapping Computer Vision Center Dept. Ciències de la Computació UAB Ricardo Toledo Morales (CVC)
A Statistical Approach to Texture Classification Nicholas Chan Heather Dunlop Project Dec. 14, 2005.
Cues We Use To Infer Depth Sensation and Perception.
Machine Vision Edge Detection Techniques ENT 273 Lecture 6 Hema C.R.
Edge Segmentation in Computer Images CSE350/ Sep 03.
Digital Image Processing CSC331
Representation in Vision Derek Hoiem CS 598, Spring 2009 Jan 22, 2009.
1 Review and Summary We have covered a LOT of material, spending more time and more detail on 2D image segmentation and analysis, but hopefully giving.
Recognizing specific objects Matching with SIFT Original suggestion Lowe, 1999,2004.
Learning Image Statistics for Bayesian Tracking Hedvig Sidenbladh KTH, Sweden Michael Black Brown University, RI, USA
Another Example: Circle Detection
MAN-522 Computer Vision Spring
Processing visual information for Computer Vision
SIFT Scale-Invariant Feature Transform David Lowe
CS262: Computer Vision Lect 09: SIFT Descriptors
A Novel 2D-to-3D Conversion System Using Edge Information
Chapter 10 Image Segmentation
Lecture 07 13/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
TP12 - Local features: detection and description
Common Classification Tasks
CAP 5415 Computer Vision Fall 2012 Dr. Mubarak Shah Lecture-5
Computer Vision Lecture 16: Texture II
Probabilistic Robotics
Grouping/Segmentation
Feature descriptors and matching
Presentation transcript:

3-D Depth Reconstruction from a Single Still Image 何開暘

Visual Cues for Depth Perception Monocular Cues  Texture variations, texture gradients, interposition, occlusion, known object sizes, light and shading, haze, defocus Stereo Cues Motion Parallax and Focus Cues

image → feature → depth Chose features that capture 3 types of cues: texture variations, texture gradients, and color Model conditional distribution of depths given monocular image features p(d|x) Estimate parameters by maximizing conditional log likelihood of training data Given an image, find MAP estimate of depths

Outline Introduction Feature Vector Probabilistic Model Experiments Reference

Feature vectors Two types of features  Absolute depth features―used to estimate absolute depth at a particular patch  Relative features―used to estimate relative depths Capture three types of cues  Texture variation―apply Law ’ s masks to intensity channel  Haze―apply a local averaging filter to color channels  Texture gradient―apply six oriented edge filters to intensity channel

Features for Absolute Depth Compute summary statistics of a patch i in the image I(x,y) as follows  Use the output of each of the 17 (9 Law’s masks, 2 color channels and 6 texture gradients) filters Fn, n=1,…,17 as: (dimension 34) To estimate absolute depth at a patch, local image features centered on the patch are insufficient Use more global properties

More Global Properties Use image features extracted at multiple spatial scales (three scale) Features used to predict depth of a particular patch are computed from that patch as well as 4 neighboring patches (Repeated at each of the three scales) Add to features of a patch additional summary features of the column it lies in (5*3+4)*34=636 dimensional

Features for Relative Depth To learn the dependencies between two neighboring patches Compute a 10-bin histogram of each of the 17 filter outputs, giving a total of 170 features y is for each patch i at scale s Relative depth features y ijs for two neighboring patches i and j at scale s will be the differences between their histogram, i.e., y ijs =y is -y js

Outline Introduction Feature Vector Probabilistic Model Experiments Reference

Gaussian Model

Laplacian Model

Outline Introduction Feature Vector Probabilistic Model Experiments Reference

Result

Improving Performance of Stereovision using Monocular Cues

The average errors as a function of the distance from the camera

Reference A.Y. Ng A. Saxena, S.H. Chung. 3-d depth reconstruction from a single still image. In International Journal of Computer Vision (IJCV), Michels, J., Saxena, A., & Ng, A. Y. (2005). High speed obstacle avoidance using monocular vision and reinforcement learning. In 22 nd international conference on machine learning (ICML).