Low-level Motion Activity Features for Semantic Characterization of Video Kadir A. Peker, A. Aydin Alatan, Ali N. Akansu International Conference on Multimedia.

Slides:



Advertisements
Similar presentations
Image Retrieval With Relevant Feedback Hayati Cam & Ozge Cavus IMAGE RETRIEVAL WITH RELEVANCE FEEDBACK Hayati CAM Ozge CAVUS.
Advertisements

Trajectory Analysis of Broadcast Soccer Videos Computer Science and Engineering Department Indian Institute of Technology, Kharagpur by Prof. Jayanta Mukherjee.
Caroline Rougier, Jean Meunier, Alain St-Arnaud, and Jacqueline Rousseau IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 21, NO. 5,
Automatic Video Shot Detection from MPEG Bit Stream Jianping Fan Department of Computer Science University of North Carolina at Charlotte Charlotte, NC.
SmartPlayer: User-Centric Video Fast-Forwarding K.-Y. Cheng, S.-J. Luo, B.-Y. Chen, and H.-H. Chu ACM CHI 2009 (international conference on Human factors.
Automatic Soccer Video Analysis and Summarization
A Novel Scheme for Video Similarity Detection Chu-Hong Hoi, Steven March 5, 2003.
Broadcast News Parsing Using Visual Cues: A Robust Face Detection Approach Yannis Avrithis, Nicolas Tsapatsoulis and Stefanos Kollias Image, Video & Multimedia.
Content-based Video Indexing, Classification & Retrieval Presented by HOI, Chu Hong Nov. 27, 2002.
ICIP 2000, Vancouver, Canada IVML, ECE, NTUA Face Detection: Is it only for Face Recognition?  A few years earlier  Face Detection Face Recognition 
ICME 2008 Huiying Liu, Shuqiang Jiang, Qingming Huang, Changsheng Xu.
1 CS 430: Information Discovery Lecture 22 Non-Textual Materials 2.
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman ICCV 2003 Presented by: Indriyati Atmosukarto.
Multimedia Search and Retrieval Presented by: Reza Aghaee For Multimedia Course(CMPT820) Simon Fraser University March.2005 Shih-Fu Chang, Qian Huang,
Efficient MPEG Compressed Video Analysis Using Macroblock Type Information Soo-Chang Pei, Yu-Zuong Chou IEEE TRANSACTIONS ON MULTIMEDIA, DECEMBER,1999.
LYU 0102 : XML for Interoperable Digital Video Library Recent years, rapid increase in the usage of multimedia information, Recent years, rapid increase.
On the Use of Computable Features for Film Classification Zeeshan Rasheed,Yaser Sheikh Mubarak Shah IEEE TRANSCATION ON CIRCUITS AND SYSTEMS FOR VIDEO.
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman University of Oxford ICCV 2003.
MPEG-7 Motion Descriptors. Reference ISO/IEC JTC1/SC29/WG11 N4031 ISO/IEC JTC1/SC29/WG11 N4062 MPEG-7 Visual Motion Descriptors (IEEE Transactions on.
CS 376b Introduction to Computer Vision 04 / 01 / 2008 Instructor: Michael Eckmann.
1 Motion in 2D image sequences Definitely used in human vision Object detection and tracking Navigation and obstacle avoidance Analysis of actions or.
Jacinto C. Nascimento, Member, IEEE, and Jorge S. Marques
Video Trails: Representing and Visualizing Structure in Video Sequences Vikrant Kobla David Doermann Christos Faloutsos.
Support Vector Machine based Logo Detection in Broadcast Soccer Videos Hossam M. Zawbaa Cairo University, Faculty of Computers and Information; ABO Research.
김덕주 (Duck Ju Kim). Problems What is the objective of content-based video analysis? Why supervised identification has limitation? Why should use integrated.
 Shutter speed is the amount of time the shutter is open while light is coming in.  Shutter speed controls whether your picture will be completely blurred,
DVMM Lab, Columbia UniversityVideo Event Recognition Video Event Recognition: Multilevel Pyramid Matching Dong Xu and Shih-Fu Chang Digital Video and Multimedia.
1 Image Video & Multimedia Systems Laboratory Multimedia Knowledge Laboratory Informatics and Telematics Institute Exploitation of knowledge in video recordings.
Utilizing Video Ontology for Fast and Accurate Query-by-Example Retrieval Kimiaki Shirahama Graduate School of Economics, Kobe University Kuniaki Uehara.
TEMPORAL VIDEO BOUNDARIES -PART ONE- SNUEE KIM KYUNGMIN.
Video Classification By: Maryam S. Mirian
Player Action Recognition in Broadcast Tennis Video with Applications to Semantic Analysis of Sport Game Guangyu Zhu, Changsheng Xu Qingming Huang, Wen.
Università degli Studi di Modena and Reggio Emilia Dipartimento di Ingegneria dell’Informazione Prototypes selection with.
Characterizing activity in video shots based on salient points Nicolas Moënne-Loccoz Viper group Computer vision & multimedia laboratory University of.
Information Systems & Semantic Web University of Koblenz ▪ Landau, Germany Semantic Web - Multimedia Annotation – Steffen Staab
An Architecture for Mining Resources Complementary to Audio-Visual Streams J. Nemrava, P. Buitelaar, N. Simou, D. Sadlier, V. Svátek, T. Declerck, A. Cobet,
Using Video The science of video Why do we perceive it? The psychology of video Where do we use it? The technology of video How can you capture and edit.
Texture. Texture is an innate property of all surfaces (clouds, trees, bricks, hair etc…). It refers to visual patterns of homogeneity and does not result.
Understanding the Semantics of Media Lecture Notes on Video Search & Mining, Spring 2012 Presented by Jun Hee Yoo Biointelligence Laboratory School of.
Performance Characterization of Video-Shot-Change Detection Methods U. Gargi, R. Kasturi, S. Strayer Presented by: Isaac Gerg.
1 CS 430: Information Discovery Lecture 22 Non-Textual Materials: Informedia.
PSEUDO-RELEVANCE FEEDBACK FOR MULTIMEDIA RETRIEVAL Seo Seok Jun.
Case Study 1 Semantic Analysis of Soccer Video Using Dynamic Bayesian Network C.-L Huang, et al. IEEE Transactions on Multimedia, vol. 8, no. 4, 2006 Fuzzy.
1 Applications of video-content analysis and retrieval IEEE Multimedia Magazine 2002 JUL-SEP Reporter: 林浩棟.
Making great productions takes more than great ideas. You need the right raw material. The storyboard can be used as a reminder of the productions content.
Digital Libraries1 David Rashty. Digital Libraries2 “A library is an arsenal of liberty” Anonymous.
Bachelor of Engineering In Image Processing Techniques For Video Content Extraction Submitted to the faculty of Engineering North Maharashtra University,
MMDB-9 J. Teuhola Standardization: MPEG-7 “Multimedia Content Description Interface” Standard for describing multimedia content (metadata).
Content-Based Image Retrieval QBIC Homepage The State Hermitage Museum db2www/qbicSearch.mac/qbic?selLang=English.
Semantic Extraction and Semantics-Based Annotation and Retrieval for Video Databases Authors: Yan Liu & Fei Li Department of Computer Science Columbia.
Image and Video Retrieval INST 734 Doug Oard Module 13.
Digital Video Library Network Supervisor: Prof. Michael Lyu Student: Ma Chak Kei, Jacky.
Colour and Texture. Extract 3-D information Using Vision Extract 3-D information for performing certain tasks such as manipulation, navigation, and recognition.
Text From Corners: A Novel Approach to Detect Text and Caption in Videos Xu Zhao, Kai-Hsiang Lin, Yun Fu, Member, IEEE, Yuxiao Hu, Member, IEEE, Yuncai.
Video Databases What are it uses? –Sports –Surveillance How do we query it? –Mosaic-based Query Language.
Learning video saliency from human gaze using candidate selection CVPR2013 Poster.
Production Techniques Equipment. Camera Shots The camera, alone or combined with narration can create the “who, what, why, where” effect we desire for.
Image features and properties. Image content representation The simplest representation of an image pattern is to list image pixels, one after the other.
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman University of Oxford ICCV 2003.
Similarity Measurement and Detection of Video Sequences Chu-Hong HOI Supervisor: Prof. Michael R. LYU Marker: Prof. Yiu Sang MOON 25 April, 2003 Dept.
Ontology-based Automatic Video Annotation Technique in Smart TV Environment Jin-Woo Jeong, Hyun-Ki Hong, and Dong-Ho Lee IEEE Transactions on Consumer.
Visual Information Retrieval
Automatic Video Shot Detection from MPEG Bit Stream
Presenter: Ibrahim A. Zedan
Saliency-guided Video Classification via Adaptively weighted learning
Using the camera Shooting etc..
Video Google: Text Retrieval Approach to Object Matching in Videos
Video Vocabulary Illustrated
Multimedia Information Retrieval
Video Google: Text Retrieval Approach to Object Matching in Videos
Presentation transcript:

Low-level Motion Activity Features for Semantic Characterization of Video Kadir A. Peker, A. Aydin Alatan, Ali N. Akansu International Conference on Multimedia and Expo 2000

Introduction We want to establish connections between low-level motion activity feature of video segments and the semantic meaningful characterization of them. Two computationally simple descriptors for motion activity of a video content is used.

Motion Activity Descriptors act 0 : monotonous (steady) motion activity descriptor act 1 : non-monotonous (unsteady) motion activity descriptor

j

act 0 is sensitive to global motion such as camera pan and to objects moving very close to camera. act 1 filters out the component of motion activity that does not change from frame to frame. In contrast to act 0, act 1 is more sensitive to unsteady motion such as fickle motion of a non-rigid object in close up.

Results from Application Examples We use the two descriptors in two different application contexts. Browsing through a sports video. Retrieval from a database of shots.

Detecting Close-ups in Sports Video We observe that the difference act 1 (n)- act 0 (n) is highest for close-up shots where the irregular motion of players in view is dominant over the regular global motion. Basketball from MPEG-7 data set (10 minutes, frames,4800 P frames) A ground truth data is prepared manually, segmenting the video into wide angle and close- up shots.(59 segments, 30 being close-ups)

We expect m1 (act 0 (n)) to be high for close-up frames because zoom or when the action is close to the camera the motion vector is larger. We expect if non-monotonous activity act 1 (n) is significantly higher than act 0 (n) in a frame, then with a high probability, the frame is a close-up on a highly active object.

Frame-based detection Two threshold for m1 and m2 to select 250 P-frames. Bounding boxes are close-up segments. Positive impulses are where m2 suggest a close-up. Negative impulses are where m1 suggest a close-up.

Segment-based Detection

We find the close-up segments by sorting the segments with respect to sm1 and sm2 and choosing the top K. The retrieval using sm1 (average of act 0 over the segment) is misled by camera motion.The first retrieved segment is a fast pan segment. We find sm2 to be a more reliable detector for close-ups.

Retrieval of High Activity Shots A database of 600 shots from MPEG-7 test set, include various programs such as news, sports,entertainment, education, etc. 5 highest activity shots are retrieved using act 0, act 1 and (act 1 - act 0 ). act 0 and act 1 retrieve shots that contain fast camera motions or an objects that passes too close to the camera, which are not commonly considered high activity. (act 1 - act 0 ) get 5 shots of dancing people.

Conclusion We described two descriptors to infer whether the activity content is dominantly a monotonous, steady motion or an unsteady, inconstant motion. This kind of a characterization of the activity content can be used to detect close-up segment in a sports video or in an activity based query from a database of video shots.