Università degli Studi di Modena and Reggio Emilia Dipartimento di Ingegneria dell’Informazione Prototypes selection with.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
DONG XU, MEMBER, IEEE, AND SHIH-FU CHANG, FELLOW, IEEE Video Event Recognition Using Kernel Methods with Multilevel Temporal Alignment.
RGB-D object recognition and localization with clutter and occlusions Federico Tombari, Samuele Salti, Luigi Di Stefano Computer Vision Lab – University.
Finding Structure in Home Videos by Probabilistic Hierarchical Clustering Daniel Gatica-Perez, Alexander Loui, and Ming-Ting Sun.
Proposed concepts illustrated well on sets of face images extracted from video: Face texture and surface are smooth, constraining them to a manifold Recognition.
Patch to the Future: Unsupervised Visual Prediction
Activity Recognition Aneeq Zia. Agenda What is activity recognition Typical methods used for action recognition “Evaluation of local spatio-temporal features.
Texture Segmentation Based on Voting of Blocks, Bayesian Flooding and Region Merging C. Panagiotakis (1), I. Grinias (2) and G. Tziritas (3)
Image Indexing and Retrieval using Moment Invariants Imran Ahmad School of Computer Science University of Windsor – Canada.
Watching Unlabeled Video Helps Learn New Human Actions from Very Few Labeled Snapshots Chao-Yeh Chen and Kristen Grauman University of Texas at Austin.
Personalized Abstraction of Broadcasted American Football Video by Highlight Selection Noboru Babaguchi (Professor at Osaka Univ.) Yoshihiko Kawai and.
A presentation by Modupe Omueti For CMPT 820:Multimedia Systems
ICME 2008 Huiying Liu, Shuqiang Jiang, Qingming Huang, Changsheng Xu.
1 Content Based Image Retrieval Using MPEG-7 Dominant Color Descriptor Student: Mr. Ka-Man Wong Supervisor: Dr. Lai-Man Po MPhil Examination Department.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
ADVISE: Advanced Digital Video Information Segmentation Engine
Segmentation Divide the image into segments. Each segment:
T.Sharon 1 Internet Resources Discovery (IRD) Introduction to MMIR.
Visual Querying By Color Perceptive Regions Alberto del Bimbo, M. Mugnaini, P. Pala, and F. Turco University of Florence, Italy Pattern Recognition, 1998.
Architecture & Data Management of XML-Based Digital Video Library System Jacky C.K. Ma Michael R. Lyu.
Video summarization by graph optimization Lu Shi Oct. 7, 2003.
Visual Information Retrieval Chapter 1 Introduction Alberto Del Bimbo Dipartimento di Sistemi e Informatica Universita di Firenze Firenze, Italy.
1 An Empirical Study on Large-Scale Content-Based Image Retrieval Group Meeting Presented by Wyman
Tracking Video Objects in Cluttered Background
Presented by Zeehasham Rasheed
CS292 Computational Vision and Language Visual Features - Colour and Texture.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
A fuzzy video content representation for video summarization and content-based retrieval Anastasios D. Doulamis, Nikolaos D. Doulamis, Stefanos D. Kollias.
DVMM Lab, Columbia UniversityVideo Event Recognition Video Event Recognition: Multilevel Pyramid Matching Dong Xu and Shih-Fu Chang Digital Video and Multimedia.
DOG I : an Annotation System for Images of Dog Breeds Antonis Dimas Pyrros Koletsis Euripides Petrakis Intelligent Systems Laboratory Technical University.
AdvisorStudent Dr. Jia Li Shaojun Liu Dept. of Computer Science and Engineering, Oakland University 3D Shape Classification Using Conformal Mapping In.
TEMPORAL VIDEO BOUNDARIES -PART ONE- SNUEE KIM KYUNGMIN.
1 Template-Based Classification Method for Chinese Character Recognition Presenter: Tienwei Tsai Department of Informaiton Management, Chihlee Institute.
MediaEval Workshop 2011 Pisa, Italy 1-2 September 2011.
Copyright R. Weber Machine Learning, Data Mining ISYS370 Dr. R. Weber.
The MPEG-7 Color Descriptors
Learning Object Metadata Mining Masoud Makrehchi Supervisor: Prof. Mohamed Kamel.
Similarity measuress Laboratory of Image Analysis for Computer Vision and Multimedia Università di Modena e Reggio Emilia,
Presented by Tienwei Tsai July, 2005
Information Systems & Semantic Web University of Koblenz ▪ Landau, Germany Semantic Web - Multimedia Annotation – Steffen Staab
General Tensor Discriminant Analysis and Gabor Features for Gait Recognition by D. Tao, X. Li, and J. Maybank, TPAMI 2007 Presented by Iulian Pruteanu.
Marcin Marszałek, Ivan Laptev, Cordelia Schmid Computer Vision and Pattern Recognition, CVPR Actions in Context.
Bayesian Extension to the Language Model for Ad Hoc Information Retrieval Hugo Zaragoza, Djoerd Hiemstra, Michael Tipping Presented by Chen Yi-Ting.
COLOR HISTOGRAM AND DISCRETE COSINE TRANSFORM FOR COLOR IMAGE RETRIEVAL Presented by 2006/8.
Pseudo-supervised Clustering for Text Documents Marco Maggini, Leonardo Rigutini, Marco Turchi Dipartimento di Ingegneria dell’Informazione Università.
Discriminative Local Binary Patterns for Human Detection in Personal Album.
Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning.
2005/12/021 Content-Based Image Retrieval Using Grey Relational Analysis Dept. of Computer Engineering Tatung University Presenter: Tienwei Tsai ( 蔡殿偉.
Event retrieval in large video collections with circulant temporal encoding CVPR 2013 Oral.
2005/12/021 Fast Image Retrieval Using Low Frequency DCT Coefficients Dept. of Computer Engineering Tatung University Presenter: Yo-Ping Huang ( 黃有評 )
Probabilistic Latent Query Analysis for Combining Multiple Retrieval Sources Rong Yan Alexander G. Hauptmann School of Computer Science Carnegie Mellon.
Summarizing Encyclopedic Term Descriptions on the Web from Coling 2004 Atsushi Fujii and Tetsuya Ishikawa Graduate School of Library, Information and Media.
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
Automatic Video Tagging using Content Redundancy Stefan Siersdorfer 1, Jose San Pedro 2, Mark Sanderson 2 1 L3S Research Center, Germany 2 University of.
Digital Video Library Network Supervisor: Prof. Michael Lyu Student: Ma Chak Kei, Jacky.
Combining Text and Image Queries at ImageCLEF2005: A Corpus-Based Relevance-Feedback Approach Yih-Cheng Chang Department of Computer Science and Information.
Categorical Perception 강우현. Introduction Scalable representations for visual categorization Representation for functional and affordance-based categorization.
MMM2005The Chinese University of Hong Kong MMM2005 The Chinese University of Hong Kong 1 Video Summarization Using Mutual Reinforcement Principle and Shot.
Similarity Measurement and Detection of Video Sequences Chu-Hong HOI Supervisor: Prof. Michael R. LYU Marker: Prof. Yiu Sang MOON 25 April, 2003 Dept.
Non-parametric Methods for Clustering Continuous and Categorical Data Steven X. Wang Dept. of Math. and Stat. York University May 13, 2010.
May 2003 SUT Color image segmentation – an innovative approach Amin Fazel May 2003 Sharif University of Technology Course Presentation base on a paper.
Ontology-based Automatic Video Annotation Technique in Smart TV Environment Jin-Woo Jeong, Hyun-Ki Hong, and Dong-Ho Lee IEEE Transactions on Consumer.
Methods of multivariate analysis Ing. Jozef Palkovič, PhD.
南台科技大學 資訊工程系 Region partition and feature matching based color recognition of tongue image 指導教授:李育強 報告者 :楊智雁 日期 : 2010/04/19 Pattern Recognition Letters,
Detecting Moving Objects, Ghosts, and Shadows in Video Streams
Working meeting of WP4 Task WP4.1
A. M. R. R. Bandara & L. Ranathunga
Visual Information Retrieval
Presenter: Ibrahim A. Zedan
Presentation transcript:

Università degli Studi di Modena and Reggio Emilia Dipartimento di Ingegneria dell’Informazione Prototypes selection with context based intra-class clustering for video annotation with MPEG-7 features Costantino Grana, Roberto Vezzani, Rita Cucchiara DELOS Conference Tirrenia (PI), Italy, February 2007

Prototypes selection with context based… 2 INTRODUCTION A system for the semi automatic annotation of videos, by means of Pictorially Enriched Ontologies is presented. These are ontologies for context-based video digital libraries, enriched by pictorial concepts for video annotation, summarization and similarity-based retrieval. Extraction of pictorial concepts with video clips clustering, ontology storing with MPEG-7, and the use of the ontology for stored video annotation are described. Video Shot & Sub-Shot detection Extraction Features Classification with ontology Save results in Mpeg 7 camera car

Prototypes selection with context based… 3 SIMILARITY OF VIDEO CLIPS 1.The color histogram, in 256 bins of HSV color space (Scalable Color Descriptor). 2.The 64 spatial color distributions: to account for the spatial distribution of the colors, an 8x8 grid is superimposed to the frame and the mean YCbCr color is computed for each area (Color Layout Descriptor). 3.The four main motion vectors: they are computed as the average of the MPEG motion vector, extracted in each quarter of frame. The median value has been adopted since MPEG motion vector are not always reliable and are often affected by noise.

Prototypes selection with context based… 4 SIMILARITY OF VIDEO CLIPS Generalization of image similarity Usually single key frame per shot, but variation may be too large. Our approach: 1. M representative frames uniformly selected from the shot 2.Remove the worse match to improve reliable correspondences

Prototypes selection with context based… 5 PICTORIAL ENRICHED ONTOLOGY CREATION 1.After the definition of the textual domain ontology, a pictorial enriched ontology requires the selection of the prototypal clips that can constitute pictorial concepts as specialization of each ontology category. Large training sets of clips for each category must be defined and an automatic process extracts some visual prototypes for every category. 2.Using the previously defined features and dissimilarity function, we employ a hierarchical clustering method, based on Complete Link. 3.This technique guarantees that each clip must be similar to every other in the cluster and any other clip outside the cluster has dissimilarity greater than the maximum distance between cluster elements. 4.For this clustering method the dissimilarity between two clusters is defined as

Prototypes selection with context based… 6

7 Automatic clustering level selection A rule has to be based on cluster topology concerns: a trade-of between data representation and small number of clusters It is not possible to choose the right one  even if we can define an objective function, the data may badly fit with that definition, in real cases. An example is provided by our experience with the Dunn’s Separation Index, which was conceived for this particular clustering approach. Better results (in terms of a subjective evaluation) have been obtained with the following approach: start from the definition of diameter and delta distance We can obtain from these the corresponding maximum and minimum at level n.

Prototypes selection with context based… 8 Automatic clustering level selection We define the Clustering Score at level n as: The selected level is the one which maximizes the clustering score. (Note that the paper says “minimize”… sigh) Note that both the maximum diameter and the minimum delta distance between clusters are monotonically increasing, so we can stop the clustering when a local minimum is reached (leading to a computational improvement).

Prototypes selection with context based… 9 Intra-class clustering with context data The presented choice of prototypes is guided by how similar the original clips are in the feature space, without considering the elements belonging to the other classes (context data). This may lead to a prototype selection which is indeed representative of the class but lacks the properties useful for discrimination purposes.

Prototypes selection with context based… 10 Intra-class clustering with context data We define an isolation coefficient for each clip as Then we can introduce a class based dissimilarity measure between two clips as: Even if the central points (B,C) are closer each other than to the corresponding colored ones (A and D respectively), the interposed purple distribution largely increases their dissimilarity measure, preventing their merge in a single cluster.

Prototypes selection with context based… 11 Intra-class clustering with context data ABBCCD distance dissimilarity

Prototypes selection with context based… 12 ONTOLOGIES IN MPEG-7 Ontologies may be effectively defined with OWL, but this language does not contain any construct for including a pictorial representation. On the other hand, such feature is present in the MPEG-7 standard. MPEG-7 has much less sophisticated tools for knowledge representation, since its purpose of standardization limits the definition of new data types, concepts and complex structures. Nevertheless, the MPEG-7 standard can naturally include pictorial elements such as objects, key-frames, clips and visual descriptors in the ontology description. Therefore, our system stores the pict-en ontology following the directions of the MPEG-7 standard, and in particular uses a double description provided by the combined with a which includes a Collection ModelType DS ClassificationSchemeDescriptionType DS …………………………… …………………………… ModelDescriptionType DS

Prototypes selection with context based… 13 Example results SkiBobF1 #Training set #Test set # of Visual Concepts NN CL CBCL Results on training set NN300 (100%) 500 (100%) CL 299 (99.7%)292 (97.3%)478 (95.6%) CBCL300 (100%)285 (95%)492 (98.4%) Results on test set NN660 (72.4%) 854 (75.8%)1181 (64.2%) CL654 (71.7%) 846 (75.1%)1159 (63.0%) CBCL657 (72%)852 (75.7%)1209 (65.7%)

Prototypes selection with context based… 14 Semi automatic annotation analysis (annotatate and correct)

Prototypes selection with context based… 15

Prototypes selection with context based… 16 Screenshot of the classification scheme manager window of the semi-automatic annotation framework, showing the classification scheme together with the selected prototypes.

Prototypes selection with context based… 17 Conclusions and future work We presented a system for the creation of a specific domain ontology, enriched with visual features and references to multimedia objects. The ontology is stored in MPEG-7 compliant format, and can be used to annotate new videos. This approach allows a system to behave differently by simply providing a different ontology, thus expanding its applicability to mixed sources Digital Libraries. We are working on the extension of the similarity measures from key frames to clip matching, by means of the Mallows distance (a linear programming based distance). Other features are being considered, and extensive tests are performed, in order to assess the scalability of the proposed approach.