Golnaz Abdollahian, Cuneyt M. Taskiran, Zygmunt Pizlo, and Edward J. Delp C AMERA M OTION -B ASED A NALYSIS OF U SER G ENERATED V IDEO IEEE TRANSACTIONS.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Road-Sign Detection and Recognition Based on Support Vector Machines Saturnino, Sergio et al. Yunjia Man ECG 782 Dr. Brendan.
電腦視覺 Computer and Robot Vision I
Change Detection C. Stauffer and W.E.L. Grimson, “Learning patterns of activity using real time tracking,” IEEE Trans. On PAMI, 22(8): , Aug 2000.
Extracting Minimalistic Corridor Geometry from Low-Resolution Images Yinxiao Li, Vidya, N. Murali, and Stanley T. Birchfield Department of Electrical and.
MPEG-4 Objective Standardize algorithms for audiovisual coding in multimedia applications allowing for Interactivity High compression Scalability of audio.
Foreground Modeling The Shape of Things that Came Nathan Jacobs Advisor: Robert Pless Computer Science Washington University in St. Louis.
Texture Segmentation Based on Voting of Blocks, Bayesian Flooding and Region Merging C. Panagiotakis (1), I. Grinias (2) and G. Tziritas (3)
A presentation by Modupe Omueti For CMPT 820:Multimedia Systems
Robust Moving Object Detection & Categorization using self- improving classifiers Omar Javed, Saad Ali & Mubarak Shah.
Haojie Li Jinhui Tang Si Wu Yongdong Zhang Shouxun Lin Automatic Detection and Analysis of Player Action in Moving Background Sports Video Sequences IEEE.
ICME 2008 Huiying Liu, Shuqiang Jiang, Qingming Huang, Changsheng Xu.
Learning to Detect A Salient Object Reporter: 鄭綱 (3/2)
Ai-Mei Huang and Truong Nguyen Video Processing LabECE Dept, UCSD, La Jolla, CA This paper appears in: Image Processing, ICIP IEEE International.
EE 7730 Image Segmentation.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Efficient Moving Object Segmentation Algorithm Using Background Registration Technique Shao-Yi Chien, Shyh-Yih Ma, and Liang-Gee Chen, Fellow, IEEE Hsin-Hua.
3. Introduction to Digital Image Analysis
Robust Object Segmentation Using Adaptive Thresholding Xiaxi Huang and Nikolaos V. Boulgouris International Conference on Image Processing 2007.
Multiple Human Objects Tracking in Crowded Scenes Yao-Te Tsai, Huang-Chia Shih, and Chung-Lin Huang Dept. of EE, NTHU International Conference on Pattern.
A Robust Scene-Change Detection Method for Video Segmentation Chung-Lin Huang and Bing-Yao Liao IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY.
Tracking Video Objects in Cluttered Background
Smart Traveller with Visual Translator. What is Smart Traveller? Mobile Device which is convenience for a traveller to carry Mobile Device which is convenience.
A Novel 2D To 3D Image Technique Based On Object- Oriented Conversion.
A REAL-TIME VIDEO OBJECT SEGMENTATION ALGORITHM BASED ON CHANGE DETECTION AND BACKGROUND UPDATING 楊靜杰 95/5/18.
A fuzzy video content representation for video summarization and content-based retrieval Anastasios D. Doulamis, Nikolaos D. Doulamis, Stefanos D. Kollias.
Object Tracking for Retrieval Application in MPEG-2 Lorenzo Favalli, Alessandro Mecocci, Fulvio Moschetti IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR.
Jacinto C. Nascimento, Member, IEEE, and Jorge S. Marques
E.G.M. PetrakisBinary Image Processing1 Binary Image Analysis Segmentation produces homogenous regions –each region has uniform gray-level –each region.
Entropy and some applications in image processing Neucimar J. Leite Institute of Computing
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Flow Based Action Recognition Papers to discuss: The Representation and Recognition of Action Using Temporal Templates (Bobbick & Davis 2001) Recognizing.
Information Extraction from Cricket Videos Syed Ahsan Ishtiaque Kumar Srijan.
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
A Local Adaptive Approach for Dense Stereo Matching in Architectural Scene Reconstruction C. Stentoumis 1, L. Grammatikopoulos 2, I. Kalisperakis 2, E.
ENT 273 Object Recognition and Feature Detection Hema C.R.
Blind Pattern Matching Attack on Watermark Systems D. Kirovski and F. A. P. Petitcolas IEEE Transactions on Signal Processing, VOL. 51, NO. 4, April 2003.
Texture. Texture is an innate property of all surfaces (clouds, trees, bricks, hair etc…). It refers to visual patterns of homogeneity and does not result.
1 Webcam Mouse Using Face and Eye Tracking in Various Illumination Environments Yuan-Pin Lin et al. Proceedings of the 2005 IEEE Y.S. Lee.
Motion Segmentation By Hadas Shahar (and John Y.A.Wang, and Edward H. Adelson, and Wikipedia and YouTube) 1.
Data Extraction using Image Similarity CIS 601 Image Processing Ajay Kumar Yadav.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
Limitations of Cotemporary Classification Algorithms Major limitations of classification algorithms like Adaboost, SVMs, or Naïve Bayes include, Requirement.
Case Study 1 Semantic Analysis of Soccer Video Using Dynamic Bayesian Network C.-L Huang, et al. IEEE Transactions on Multimedia, vol. 8, no. 4, 2006 Fuzzy.
Region-Based Saliency Detection and Its Application in Object Recognition IEEE TRANSACTIONS ON CIRCUITS AND SYSTEM FOR VIDEO TECHNOLOGY, VOL. 24 NO. 5,
CVPR2013 Poster Detecting and Naming Actors in Movies using Generative Appearance Models.
Image-Based Segmentation of Indoor Corridor Floors for a Mobile Robot Yinxiao Li and Stanley T. Birchfield The Holcombe Department of Electrical and Computer.
Jiu XU, Axel BEAUGENDRE and Satoshi GOTO Computer Sciences and Convergence Information Technology (ICCIT), th International Conference on 1 Real-time.
Segmentation of Vehicles in Traffic Video Tun-Yu Chiang Wilson Lau.
 Present by 陳群元.  Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental.
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
Text From Corners: A Novel Approach to Detect Text and Caption in Videos Xu Zhao, Kai-Hsiang Lin, Yun Fu, Member, IEEE, Yuxiao Hu, Member, IEEE, Yuncai.
Journal of Visual Communication and Image Representation
Vehicle Detection in Aerial Surveillance Using Dynamic Bayesian Networks Hsu-Yung Cheng, Member, IEEE, Chih-Chia Weng, and Yi-Ying Chen IEEE TRANSACTIONS.
Page 11/28/2016 CSE 40373/60373: Multimedia Systems Quantization  F(u, v) represents a DCT coefficient, Q(u, v) is a “quantization matrix” entry, and.
Wonjun Kim and Changick Kim, Member, IEEE
Learning video saliency from human gaze using candidate selection CVPR2013 Poster.
1/39 Motion Adaptive Search for Fast Motion Estimation 授課老師:王立洋老師 製作學生: M 蔡鐘葳.
Color Image Segmentation Mentor : Dr. Rajeev Srivastava Students: Achit Kumar Ojha Aseem Kumar Akshay Tyagi.
May 2003 SUT Color image segmentation – an innovative approach Amin Fazel May 2003 Sharif University of Technology Course Presentation base on a paper.
Motion tracking TEAM D, Project 11: Laura Gui - Timisoara Calin Garboni - Timisoara Peter Horvath - Szeged Peter Kovacs - Debrecen.
Student Gesture Recognition System in Classroom 2.0 Chiung-Yao Fang, Min-Han Kuo, Greg-C Lee, and Sei-Wang Chen Department of Computer Science and Information.
Traffic Sign Recognition Using Discriminative Local Features Andrzej Ruta, Yongmin Li, Xiaohui Liu School of Information Systems, Computing and Mathematics.
Fast and Robust Object Tracking with Adaptive Detection
A New Approach to Track Multiple Vehicles With the Combination of Robust Detection and Two Classifiers Weidong Min , Mengdan Fan, Xiaoguang Guo, and Qing.
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
EM Algorithm and its Applications
Presentation transcript:

Golnaz Abdollahian, Cuneyt M. Taskiran, Zygmunt Pizlo, and Edward J. Delp C AMERA M OTION -B ASED A NALYSIS OF U SER G ENERATED V IDEO IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 12, NO. 1, JANUARY 2010

 UGV generally has a rich camera motion structure that is generated by the person taking the video and it is typically unedited and unstructured.  The main application of our system is for mobile devices which have become more popular for recording, sharing, downloading and watching UGV  use computationally efficient methods  We propose a new location-based saliency map which uses camera motion information to determine the saliency values of pixels with respect to their spatial location in the frame. Introduction

 Global Motion Estimation In the majority of UGV, camera motion is limited to a few operations, e.g. pan, tilt, and zoom; more complex camera movements, such as rotation, rarely occur in UGV our goal here is to be computationally efficient to be able to target devices with low processing power such as mobile devices use a simplified three-parameter global camera motion model in the three major directions Motion-Based Frame Labeling H :horizontal V : vertical R : radial

 Template   The iteration stops when a local minimum is found Motion-Based Frame Labeling L1 distance between the 2-D template in the current frame and the previous template

 Motion Classification – support vector machine(SVM) is used  We first classify it as having a zoom or not, using the 3-Dmotion vector as the feature vector  SVM classifiers are trained on an eight-dimensional feature vector derived from the parameters H and V over a temporal sliding window. The size of sliding window is different for blurry(N=7) and shaky(N=31)  The frames that are not labeled as zoom, blurry or shaky are identified as stable motion with no zooms. Motion-Based Frame Labeling

 Two frames are considered to be correlated if they have overlap with each other  Camera View : a temporal concept defined as a set of consecutive frames that are all correlated with each other.  View boundaries occur when the camera is displaced or there is a change of viewing angle  To detect view boundaries for temporally segmenting the video,we defind the displacement vector between frames i and j as T emporal V ideo S egmentation B ased on T he U se Of C amera V iew

 A boundary frame is flagged whenever the magnitude of the displacement vector,, for the current frame and that for the previously detected boundary frame is larger than  There is a constraint that boundary frame can’t be chosen during intervals labeled as blurry segments. T emporal V ideo S egmentation B ased on T he U se Of C amera V iew

 A keyframe should be the frame with the highest subjective importance in the segment in order to represent the segment it is extracted from  Since our intention was to avoid the complex tasks of object and action recognition in our system, our keyframe selection strategy was only based on camera motion factor.  The following frames are selected as keyframes :  The frame after a zoom-in  The frame after a large zoom-out  The frame where the camera is at pause  For segments during which the camera has constant motion, all frames are considered to be of relatively same importance. In this case, the frame closest to the middle of the segment and having the least amount of motion is chosen as the keyframe in order to minimize blurriness Keyframe Selection

Combine several saliency map to generate the keyframes saliency maps  color contrast saliency  moving objects saliency map  highlighted faces  location-based saliency map Keyframe Saliency Maps and ROI Extraction

 Use the RGB color space to generate the contrast-based saliency map  The three-dimensional pixel vectors in RGB space are clustered into a small number of color vectors using generalized Lloyd algorithm (GLA) for vector quantization  Color Contrast Saliency Map Pij and q : RGB pixel value Θ : neighborhood of pixel (i,j) (5*5) d : gaussian distance

 To determine the moving object saliency map, we examine the magnitude and phase of macro block relative motion vectors  Relative motion vector for the macro block at location (m,n) :  If relative motion below a threshold values -> assign 0  The motion intensity I and motion phase φ are defined as Moving Object Saliency Map

 The phase entropy map, H p, indicates the regions with inconsistent motion which usually belong to the boundary of the moving object  Moving Object Saliency Map the probability of the kth phase whose value is estimated from the histogram

 The direction of the camera motion also has a major effect on the regions where a viewer “looks” in the sequence  The global motion parameters were used to generate the location saliency maps for the extracted keyframes Location-Based Saliency Map k H,k V,K r : constant (10,5,0.5) r : distance of a pixel from the center r max : maximum r in the frame

 After combining the H and V maps, the peak of the map function is at  The radial map, S R, is either decreasing or increasing as we move from the center to the borders, depending on whether the camera has a zoom-in/no-zoom or zoom-out operation Location-Based Saliency Map

 First, the color contrast and moving object saliency maps are superimposed since they represent two independent factors in attracting visual attention  Faces are detected and highlighted after combining the low level saliency maps  The location-based saliency map is then multiplied pixel-wise with this map to yield the combined saliency map Combined Saliency Map

 A region growing algorithm proposed is used to extract ROI from the saliency map  Fuzzy partitioning is employed to classify the pixels into R 1 : ROI and R 0 : insignificant regions  seed selection  1) the seeds must have maximum local contrast  2) the seeds should belong to the attended areas Identification of ROIs

Experimental Results

left Zoom-out Zoom-in

Experimental Results

 UGVs contain rich content-based camera motion structure that can be an indicator of “importance” in the scene  Since camera motion in UGV may have both intentional and unintentional behaviors, we used motion classification as a preprocessing step  A temporal segmentation algorithm was proposed based on the concept of camera views which relates each subshot to a different view  We use a simple keyframe selection strategy based on camera motion patterns to represent each view  we employed camera motion in addition to several other factors to generate saliency maps for keyframes and identify ROIs based on visual attention Conclusion