Speaker: Chi-Yu Hsu Advisor: Prof. Jian-Jung Ding Leveraging Stereopsis for Saliency Analysis, CVPR 2012.

Slides:



Advertisements
Similar presentations
A Unified Framework for Context Assisted Face Clustering
Advertisements

Automatic Image Annotation Using Group Sparsity
Seeing 3D from 2D Images. How to make a 2D image appear as 3D! ► Output and input is typically 2D Images ► Yet we want to show a 3D world! ► How can we.
Ming-Ming Cheng 1 Ziming Zhang 2 Wen-Yan Lin 3 Philip H. S. Torr 1 1 Oxford University, 2 Boston University 3 Brookes Vision Group Training a generic objectness.
Leveraging Stereopsis for Saliency Analysis
Recovering Human Body Configurations: Combining Segmentation and Recognition Greg Mori, Xiaofeng Ren, and Jitentendra Malik (UC Berkeley) Alexei A. Efros.
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Patch to the Future: Unsupervised Visual Prediction
Qualifying Exam: Contour Grouping Vida Movahedi Supervisor: James Elder Supervisory Committee: Minas Spetsakis, Jeff Edmonds York University Summer 2009.
GrabCut Interactive Image (and Stereo) Segmentation Carsten Rother Vladimir Kolmogorov Andrew Blake Antonio Criminisi Geoffrey Cross [based on Siggraph.
Hierarchical Saliency Detection School of Electronic Information Engineering Tianjin University 1 Wang Bingren.
The Apparatus. Seeing in Stereo It’s very hard to read words if there are multiple images on your retina.
IEEE TCSVT 2011 Wonjun Kim Chanho Jung Changick Kim
Learning to Detect A Salient Object Reporter: 鄭綱 (3/2)
Stas Goferman Lihi Zelnik-Manor Ayellet Tal. …
Boundary matting for view synthesis Samuel W. Hasinoff Sing Bing Kang Richard Szeliski Computer Vision and Image Understanding 103 (2006) 22–32.
A Novel 2D-to-3D Conversion System Using Edge Information IEEE Transactions on Consumer Electronics 2010 Chao-Chung Cheng Chung-Te li Liang-Gee Chen.
What is Stereopsis? The process in visual perception that leads to the sensation of depth due to the slightly different perspectives that our two eyes.
CS335 Principles of Multimedia Systems Content Based Media Retrieval Hao Jiang Computer Science Department Boston College Dec. 4, 2007.
Presented by Zeehasham Rasheed
A Novel 2D To 3D Image Technique Based On Object- Oriented Conversion.
IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 11, NOVEMBER 2011 Qian Zhang, King Ngi Ngan Department of Electronic Engineering, the Chinese university.
December 2, 2014Computer Vision Lecture 21: Image Understanding 1 Today’s topic is.. Image Understanding.
Ganglion cells project to the brain via the optic nerve information is projected to contralateral cortex! Visual Pathways.
Introduction of Saliency Map
Computer Vision Spring ,-685 Instructor: S. Narasimhan WH 5409 T-R 10:30am – 11:50am Lecture #15.
Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’04) /04 $20.00 c 2004 IEEE 1 Li Hong.
Satellites in Our Pockets: An Object Positioning System using Smartphones Justin Manweiler, Puneet Jain, Romit Roy Choudhury TsungYun
Linked Edges as Stable Region Boundaries* Michael Donoser, Hayko Riemenschneider and Horst Bischof This work introduces an unsupervised method to detect.
CAP4730: Computational Structures in Computer Graphics 3D Concepts.
Mutual Information-based Stereo Matching Combined with SIFT Descriptor in Log-chromaticity Color Space Yong Seok Heo, Kyoung Mu Lee, and Sang Uk Lee.
ICPR/WDIA-2012 High Quality Novel View Synthesis Based on Low Resolution Depth Image and High Resolution Color Image Jui-Chiu Chiang, Zheng-Feng Liu, and.
Automatic Registration of Color Images to 3D Geometry Computer Graphics International 2009 Yunzhen Li and Kok-Lim Low School of Computing National University.
By Andrea Rees. Gestalt Principles 1) Closure 2) Proximity 3) Similarity 4) Figure VISUAL PERCEPTION PRINCIPLES OVERVIEW Depth Principles Binocular 1)
Object Stereo- Joint Stereo Matching and Object Segmentation Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on Michael Bleyer Vienna.
Model comparison and challenges II Compositional bias of salient object detection benchmarking Xiaodi Hou K-Lab, Computation and Neural Systems California.
User Issues in 3D TV & Cinema Martin S. Banks Vision Science Program UC Berkeley.
1 Perception, Illusion and VR HNRS 299, Spring 2008 Lecture 8 Seeing Depth.
Automatic Image Annotation by Using Concept-Sensitive Salient Objects for Image Content Representation Jianping Fan, Yuli Gao, Hangzai Luo, Guangyou Xu.
Geodesic Saliency Using Background Priors
Region-Based Saliency Detection and Its Application in Object Recognition IEEE TRANSACTIONS ON CIRCUITS AND SYSTEM FOR VIDEO TECHNOLOGY, VOL. 24 NO. 5,
CVPR2013 Poster Detecting and Naming Actors in Movies using Generative Appearance Models.
Saliency Aggregation: A Data- driven Approach Long Mai Yuzhen Niu Feng Liu Department of Computer Science, Portland State University Portland, OR,
Spatiotemporal Saliency Map of a Video Sequence in FPGA hardware David Boland Acknowledgements: Professor Peter Cheung Mr Yang Liu.
Visual Perception Principles Visual perception principles are ‘rules’ that we apply to visual information to assist our organisation and interpretation.
Stas Goferman Lihi Zelnik-Manor Ayellet Tal Technion.
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
Stereoscopic Images Binocular vision enables us to measure depth using eye convergence and stereoscopic vision. Eye convergence is a measure of the angle.
Journal of Visual Communication and Image Representation
Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth
Learning Photographic Global Tonal Adjustment with a Database of Input / Output Image Pairs.
1 Computational Vision CSCI 363, Fall 2012 Lecture 16 Stereopsis.
Grenoble Images Parole Signal Automatique Modeling of visual cortical processing to estimate binocular disparity Introduction - The objective is to estimate.
Visual Perception. What is Visual Perception? Visual perception are rules we apply to visual information to assist our organisation and interpretation.
VISUAL PERCEPTION PRINCIPLES By Mikayla. VISUAL PERCEPTION PRINCIPLES  Gestalt principles 1.Closure 2.Proximity 3.Similarity 4.Figure-ground  Depth.
A computational model of stereoscopic 3D visual saliency School of Electronic Information Engineering Tianjin University 1 Wang Bingren.
 Mentor : Prof. Amitabha Mukerjee Learning to Detect Salient Objects Team Members - Avinash Koyya Diwakar Chauhan.
Computational Vision CSCI 363, Fall 2012 Lecture 17 Stereopsis II
Minimum Barrier Salient Object Detection at 80 FPS JIANMING ZHANG, STAN SCLAROFF, ZHE LIN, XIAOHUI SHEN, BRIAN PRICE, RADOMIR MECH IEEE INTERNATIONAL CONFERENCE.
Shadow Detection in Remotely Sensed Images Based on Self-Adaptive Feature Selection Jiahang Liu, Tao Fang, and Deren Li IEEE TRANSACTIONS ON GEOSCIENCE.
Correspondence and Stereopsis. Introduction Disparity – Informally: difference between two pictures – Allows us to gain a strong sense of depth Stereopsis.
A Novel 2D-to-3D Conversion System Using Edge Information
Depth Perception, with Emphasis on Stereoscopic Vision
Computational Vision CSCI 363, Fall 2016 Lecture 15 Stereopsis
Saliency detection Donghun Yeo CV Lab..
Learning to Detect a Salient Object
Enhanced-alignment Measure for Binary Foreground Map Evaluation
Dingding Liu* Yingen Xiong† Linda Shapiro* Kari Pulli†
Saliency detection Donghun Yeo CV Lab..
Saliency Optimization from Robust Background Detection
Presentation transcript:

Speaker: Chi-Yu Hsu Advisor: Prof. Jian-Jung Ding Leveraging Stereopsis for Saliency Analysis, CVPR 2012

Outline Abstract Introduction Stereo Saliency Experiments Conclusion

Abstract

Introduction Stereopsis provides an additional depth cue and plays an important role in the human vision system. Two approaches: - computes stereo saliency based on the global disparity contrast in the input image. - leverages domain knowledge in stereoscopic photography. Build a stereo saliency analysis benchmark dataset that contains 1000 stereoscopic images with salient object masks.

Stereopsis Stereopsis is a process in visual perception leading to the perception of depth from retinal disparities. Created by projecting two slightly different images onto the retinas of the two eyes. Stereoscopic depth often enables people to identify objects from the background with similar visual attributes.

Stereo Saliency Compute stereo saliency from the disparity map between the left and right view of a stereoscopic image. Low-level disparity contrast Unique domain knowledge

Disparity Map Estimates a disparity map from the left and right view of a stereoscopic image. Dense disparity map is obtained from stereo matching. Apply the SIFT flow method to disparity estimation for its robustness.

Example

Stereo Saliency from Disparity Contrast When taking a photo, people tend to place an important object at a different depth level than the others. Extend a recent color contrast-based saliency detection method from Cheng et al (RC). Consider both color and disparity during segmentation by treating the disparity value as a fourth channel.

Stereo Saliency from Disparity Contrast The saliency value for each region is computed based on its contrast with all the others in the image. Sc(Ri): the saliency for region Ri. d(Ri,Rk): the disparity difference between Ri and Rk. (the average disparity difference between each pixel in Ri and Rk.) nk: the size of Rk.

Stereo Saliency from Disparity Contrast

Domain Knowledge Assisted Saliency Analysis A zone close to the screen is called the comfort zone. In practice, content of interest is often given small or zero disparities to minimize the vergence-accommodation conflict. Two rules to compute knowledge-based stereo saliency. 1. Objects with small disparity magnitudes (e.g. in the comfort zone) tend to be salient. 2. Objects popping out from the screen tend to be salient.

Stereoscopic Comfort Zone

Vergence A vergence is the simultaneous movement of both eyes in opposite directions to obtain or maintain single binocular vision. The two eyes converge to point to the same object.

Accommodation Accommodation (Acc) is the process by which the vertebrate eye changes optical power to maintain a clear image (focus) on an object as its distance varies.

Rule 1 Objects with small disparity magnitudes (e.g. in the comfort zone) tend to be salient. dmax: the maximal disparity. dmin: the minimal disparity. di: the average disparity in region Ri.

Rule 2 Objects popping out from the screen tend to be salient. Objects with negative disparities are perceived popping out from the screen. This suggests that the more negative an object’s disparity is, the more it is perceived popping out.

Rule 1 and Rule 2

How to combine R1 and R2 When an image only has negative disparities, the whole scene is perceived popping out from the screen. When an image has both negative and positive disparities, it is more likely that the disparity distribution of this image has been adjusted either manually or automatically by the camera to minimize the vergence-accommodation conflict. : the number of pixels with a negative disparity and n is the total number of pixels in the image. γ : a parameter with default value 0.5.

A common problem Some background regions will be detected as salient. Disparities change little in each row in background. Compute the local disparity contrast along each row. p: a pixel in region Ri. dp: its disparity :the average disparity of the row that contains p.

Local contrast-based saliency analysis

The Final Stereo Saliency Map Multiplies the global disparity contrast-based saliency Sc with the domain knowledge-based saliency Sr as the final stereo saliency.

Experiments Database There is no publicly available stereoscopic image database for saliency analysis stereoscopic images from Flickr, Stereoscopic Image Gallery, and NVIDIA 3D Vision Live. Three users are asked to enclose the most salient object in each image with a rectangle. Remove the images with the least consistent labels. Ask a user to manually segment the salient object stereoscopic images with human-labeled mask.

Benchmark Examples The first image was consistently labeled by three users and was included into our dataset. Its salient object mask is shown in the middle. The last image was not consistently labeled and was removed from our dataset.

Performance Evaluation Evaluate the three saliency maps from our methods, namely global disparity contrast-based stereo saliency (CSS), knowledge-assisted stereo saliency (KSS), and their combination (SS). Selected six state-of-the-art saliency detection methods, including RC, CA GB, FT, SR, and MS.

Visual Examples

Precision and Recall

Automatic Salient Object Segmentation Initializes the segmentation with a binarized saliency map. The threshold which gives 95% recall rate in the previous experiment The Grab-Cut method is extended to take the binarized saliency map as input and is iteratively applied to segmentation. Use both color and disparity to compute the difference between two pixels in the Grab-Cut framework.

Precision and Recall

Visual Examples

Limitations Performance of the methods depend on the quality of disparity maps. Stereo saliency is useful only if a salient object stays at a different depth than its surroundings. This is the same as saliency analysis w.r.t other visual features. If an object cannot be separated from the background w.r.t one feature, the saliency analysis based on that feature fails.

Conclusion Two methods for stereo saliency detection. The first method computes stereo saliency based on global disparity contrast. The second method makes use of the stereoscopic rules for saliency estimation. Experiments showed that stereo saliency is a useful complement to existing visual saliency analysis and the methods are able to achieve high-quality stereo saliency detection results.

Q&A

Reference Yuzhen Niu, Yujie Geng, Xueqing Li, and Feng Liu. Leveraging Stereopsis for Saliency Analysis. IEEE CVPR 2012.