Reading Notes: Special Issue on Distributed Smart Cameras, Proceedings of the IEEE Mahmut Karakaya Graduate Student Electrical Engineering and Computer.

Slides:



Advertisements
Similar presentations
Distinctive Image Features from Scale-Invariant Keypoints
Advertisements

Complex Networks for Representation and Characterization of Images For CS790g Project Bingdong Li 9/23/2009.
3D Model Matching with Viewpoint-Invariant Patches(VIP) Reporter :鄒嘉恆 Date : 10/06/2009.
V-1 Part V: Collaborative Signal Processing Akbar Sayeed.
Change Detection C. Stauffer and W.E.L. Grimson, “Learning patterns of activity using real time tracking,” IEEE Trans. On PAMI, 22(8): , Aug 2000.
Coverage Estimation in Heterogeneous Visual Sensor Networks Mahmut Karakaya and Hairong Qi Advanced Imaging & Collaborative Information Processing Laboratory.
TP14 - Local features: detection and description Computer Vision, FCUP, 2014 Miguel Coimbra Slides by Prof. Kristen Grauman.
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Object Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition l Panoramas,
Learning to estimate human pose with data driven belief propagation Gang Hua, Ming-Hsuan Yang, Ying Wu CVPR 05.
Neurocomputing,Neurocomputing, Haojie Li Jinhui Tang Yi Wang Bin Liu School of Software, Dalian University of Technology School of Computer Science,
Object Inter-Camera Tracking with non- overlapping views: A new dynamic approach Trevor Montcalm Bubaker Boufama.
Robust Object Tracking via Sparsity-based Collaborative Model
Multiple People Detection and Tracking with Occlusion Presenter: Feifei Huo Supervisor: Dr. Emile A. Hendriks Dr. A. H. J. Stijn Oomes Information and.
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
3D Computer Vision and Video Computing Review Midterm Review CSC I6716 Spring 2011 Prof. Zhigang Zhu
Real-time Embedded Face Recognition for Smart Home Fei Zuo, Student Member, IEEE, Peter H. N. de With, Senior Member, IEEE.
MULTI-TARGET TRACKING THROUGH OPPORTUNISTIC CAMERA CONTROL IN A RESOURCE CONSTRAINED MULTIMODAL SENSOR NETWORK Jayanth Nayak, Luis Gonzalez-Argueta, Bi.
A Study of Approaches for Object Recognition
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Motion based Correspondence for Distributed 3D tracking of multiple dim objects Ashok Veeraraghavan.
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2005 with a lot of slides stolen from Steve Seitz and.
Distinctive image features from scale-invariant keypoints. David G. Lowe, Int. Journal of Computer Vision, 60, 2 (2004), pp Presented by: Shalomi.
Object Recognition Using Distinctive Image Feature From Scale-Invariant Key point D. Lowe, IJCV 2004 Presenting – Anat Kaspi.
Scale Invariant Feature Transform (SIFT)
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
TelosCAM: Identifying Burglar Through Networked Sensor-Camera Mates with Privacy Protection Presented by Qixin Wang Shaojie Tang, Xiang-Yang Li, Haitao.
3-D Computer Vision Using Structured Light Prepared by Burak Borhan.
Scale-Invariant Feature Transform (SIFT) Jinxiang Chai.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography.
College of Engineering and Science Clemson University
Vision Guided Robotics
CSE 6367 Computer Vision Stereo Reconstruction Camera Coordinate Transformations “Everything should be made as simple as possible, but not simpler.” Albert.
Automatic Camera Calibration
Distinctive Image Features from Scale-Invariant Keypoints By David G. Lowe, University of British Columbia Presented by: Tim Havinga, Joël van Neerbos.
Computer vision.
Lecture 11 Stereo Reconstruction I Lecture 11 Stereo Reconstruction I Mata kuliah: T Computer Vision Tahun: 2010.
Final Exam Review CS485/685 Computer Vision Prof. Bebis.
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
Recognition and Matching based on local invariant features Cordelia Schmid INRIA, Grenoble David Lowe Univ. of British Columbia.
Object Tracking/Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition.
Distributed solutions for visual sensor networks to detect targets in crowds Cheng Qian.
Computer Vision Lab Seoul National University Keyframe-Based Real-Time Camera Tracking Young Ki BAIK Vision seminar : Mar Computer Vision Lab.
1 Collaborative Processing in Sensor Networks Lecture 5 - Visual Coverage Hairong Qi, Associate Professor Electrical Engineering and Computer Science University.
Acquiring 3D models of objects via a robotic stereo head David Virasinghe Department of Computer Science University of Adelaide Supervisors: Mike Brooks.
CSCE 643 Computer Vision: Extractions of Image Features Jinxiang Chai.
Wenqi Zhu 3D Reconstruction From Multiple Views Based on Scale-Invariant Feature Transform.
Vehicle Segmentation and Tracking From a Low-Angle Off-Axis Camera Neeraj K. Kanhere Committee members Dr. Stanley Birchfield Dr. Robert Schalkoff Dr.
Jack Pinches INFO410 & INFO350 S INFORMATION SCIENCE Computer Vision I.
FREE-VIEW WATERMARKING FOR FREE VIEW TELEVISION Alper Koz, Cevahir Çığla and A.Aydın Alatan.
COMP 417 – Jan 12 th, 2006 Guest Lecturer: David Meger Topic: Camera Networks for Robot Localization.
A Tutorial on using SIFT Presented by Jimmy Huff (Slightly modified by Josiah Yoder for Winter )
Visual Odometry David Nister, CVPR 2004
Local features: detection and description
3D Reconstruction Using Image Sequence
Stereo Vision Local Map Alignment for Robot Environment Mapping Computer Vision Center Dept. Ciències de la Computació UAB Ricardo Toledo Morales (CVC)
Recognizing specific objects Matching with SIFT Original suggestion Lowe, 1999,2004.
Motion Estimation of Moving Foreground Objects Pierre Ponce ee392j Winter March 10, 2004.
Computer vision: geometric models Md. Atiqur Rahman Ahad Based on: Computer vision: models, learning and inference. ©2011 Simon J.D. Prince.
Processing visual information for Computer Vision
Interest Points EE/CSE 576 Linda Shapiro.
SENSOR FUSION LAB RESEARCH ACTIVITIES PART I : DATA FUSION AND DISTRIBUTED SIGNAL PROCESSING IN SENSOR NETWORKS Sensor Fusion Lab, Department of Electrical.
Paper – Stephen Se, David Lowe, Jim Little
TP12 - Local features: detection and description
A Forest of Sensors: Using adaptive tracking to classify and monitor activities in a site Eric Grimson AI Lab, Massachusetts Institute of Technology
Cheng-Ming Huang, Wen-Hung Liao Department of Computer Science
Vehicle Segmentation and Tracking in the Presence of Occlusions
Recognition and Matching based on local invariant features
Presentation transcript:

Reading Notes: Special Issue on Distributed Smart Cameras, Proceedings of the IEEE Mahmut Karakaya Graduate Student Electrical Engineering and Computer Science University of Tennessee, Knoxville

Outline Introduction to Visual Sensor Networks (VSNs) Limitations of Visual Sensor Platforms Applications of VSN Challenges in Visual Sensors Trends in VSN Basic Concepts for Target Detection, Tracking and Recognition Camera Calibration with Belief Propagation Overview of papers from Proceedings of the IEEE, 96(10), October 2008 “An Introduction to Distributed Smart Cameras” by Rinner and Wolf. “Object Detection, Tracking and Recognition for MSCs” by Sankaranarayanan, Veeraraghavan, Chellappa. “Calibrating Distributed Camera Networks” by Devarajan, Cheng, Radke

Introduction to Visual Sensor Networks What is MSP?  Visual sensing  Computing  Wireless comm.  Collaborative visual computing  Centralized  Distributed What is VSN?  Collaboration in VSN:  To compensate for the limitations of each sensor node,  To improve the accuracy and robustness of the sensor network. * Figure Courtesy of [1]

Limitations of Visual Sensor Platforms SMALL SIZE, LOW POWER, LOW-COST = Low-processing speed Microprocessors and FPGA Small memory space Huge data volume i.e. image/video Bottleneck of system performance Low-bandwidth communication, Low level comm. protocols, ZigBee. $(Communication) > 1000x$(Computation) Scarce energy sources. Powered by battery.

Applications  Intelligent Video Surveillance Systems: Detect abnormal behaviors in the observed scene.  Intelligent Transportation Systems: Traffic monitoring, intelligent automobiles (inside and outside of the vehicle).  Medicine: Monitor patients, to see live retinal images during surgery.  Entertainment and Smart Environments: Gesture recognition.

Limited Field of View (FOV) Directional sensing Angle of view of normal lens 25 o -50 o. Visual Occlusion: Static and Dynamic Capture a target only when It stands in the field of view, No other occluding targets Challenges in Visual Sensors * Figure Courtesy of [1]

Trends in VSN From Static to Dynamic and Adaptive: To execute adaptive algorithms in order to better account for changes in the observed scene, To exploit static and mobile cameras such as PTZ cameras, To change camera functionality to provide increased autonomy. From Small to Very Large Camera Sets: Consider new types of information that to extract from the cameras. From Vision-Only to Multi-sensor Systems: Integrate different sensors i.e. audio, seismic, thermal Exploits the distinct characteristics of the individual sensors resulting in an enhanced overall output.

Distributed Computing & Distributed Camera Networks Distributed computing: a network of processors, not have direct knowledge of the state of other nodes. Distributed algorithms are designed to minimize the number of messages required to complete the algorithm. Processing all the data centrally poses several problems because of the data volume so process it locally before transmission. Data only go to the necessary nodes. Partition the network, so that use bandwidth efficiently. Real-time considerations: A distributed system can make sure only relevant nodes are involved in a given decision.

Basic Concepts for Target Detection and Tracking in VSN Central projection: Map the targets from real-world to image plane by using the projection matrix. Epipolar Geometry: Epipolar lines between focal point of two cameras and a pixel-point object are used to obtain correspondence across multiple views. Triangulation: The correspondences between multiple views is be utilized to localize targets. * Figure Courtesy of [2]

Features Extraction for Target Recognition in VSN Global properties: i.e. shape, color, texture Very sensitive to external conditions such as lighting, pose, viewpoint etc. Local features: i.e. discriminative points Discriminative points are chosen by using feature detector methods i.e. Harris corner detector and scale-invariant feature transform (SIFT). The object is defined by using these local descriptors which are very robust to viewpoint and illumination.

Geometry of the Camera & Parameters Typically, a camera is described by two sets of parameters:  The Internal parameters  The focal length  Position of principal points  The skew.  The external parameters describe the placement of the camera in a world coordinate system using  The rotation matrix and  The translation vector. * Figure Courtesy of Satya Prakash Mallick

Neighborhood Clustering Figure: (a) A snapshot of the instantaneous state of a camera network, indicating the fields of view of ten cameras. (b) The associated communication graph. (c) The associated vision graph Vision Graph: Detect the feature points in the image, matching the detected features through the neighbor nodes in vision graph. * Figure Courtesy of [3]

Feature Point Detection Figure: (a) Original image, with detected feature points overlaid. (b) Top view of the reconstructed 3-D scene and camera configuration for the experiment. Scale-invariant feature transform (SIFT): determines the “blobs” by processing the image at multiple scales with difference of Gaussian (DoG) filter and returning the scale-space extreme points of the result.

Belief Propagation Each camera node estimates the calibration parameters of neighboring cameras and updates them in a distributed algorithm to reach the accurate calibration. Summary of BP For each node Forms a neighborhood cluster, independently Performs the local calibration, based on common scene points. Calibrated nodes and scene points are incrementally merged into a common coordinate frame.

Questions THANK YOU…..

References [1] Rinner, B.; Wolf, W. “An Introduction to Distributed Smart Cameras” Proceedings of the IEEE, 96(10), p October 2008 [2] Sankaranarayanan, A.C.; Veeraraghavan, A.; Chellappa, R. “Object Detection, Tracking and Recognition for Multiple Smart Cameras” Proceedings of the IEEE, 96(10), p , October 2008 [3] Devarajan, D.; Zhaolin Cheng; Radke, R.J. “Calibrating Distributed Camera Networks” Proceedings of the IEEE, 96(10), p , October 2008