NN k Networks for browsing and clustering image collections Daniel Heesch Communications and Signal Processing Group Electrical and Electronic Engineering.

Slides:



Advertisements
Similar presentations
Image Retrieval With Relevant Feedback Hayati Cam & Ozge Cavus IMAGE RETRIEVAL WITH RELEVANCE FEEDBACK Hayati CAM Ozge CAVUS.
Advertisements

Complex Networks for Representation and Characterization of Images For CS790g Project Bingdong Li 9/23/2009.
INFO624 - Week 2 Models of Information Retrieval Dr. Xia Lin Associate Professor College of Information Science and Technology Drexel University.
Clustering Categorical Data The Case of Quran Verses
Carolina Galleguillos, Brian McFee, Serge Belongie, Gert Lanckriet Computer Science and Engineering Department Electrical and Computer Engineering Department.
Content Based Image Clustering and Image Retrieval Using Multiple Instance Learning Using Multiple Instance Learning Xin Chen Advisor: Chengcui Zhang Department.
Image Search Presented by: Samantha Mahindrakar Diti Gandhi.
Chapter 5: Query Operations Baeza-Yates, 1999 Modern Information Retrieval.
CONTENT BASED FACE RECOGNITION Ankur Jain 01D05007 Pranshu Sharma Prashant Baronia 01D05005 Swapnil Zarekar 01D05001 Under the guidance of Prof.
Efficient and Robust Computation of Resource Clusters in the Internet Efficient and Robust Computation of Resource Clusters in the Internet Chuang Liu,
Relevance Feedback based on Parameter Estimation of Target Distribution K. C. Sia and Irwin King Department of Computer Science & Engineering The Chinese.
February, Content-Based Image Retrieval Saint-Petersburg State University Natalia Vassilieva Il’ya Markov
1 Integrating User Feedback Log into Relevance Feedback by Coupled SVM for Content-Based Image Retrieval 9-April, 2005 Steven C. H. Hoi *, Michael R. Lyu.
Clustering Color/Intensity
Visual Querying By Color Perceptive Regions Alberto del Bimbo, M. Mugnaini, P. Pala, and F. Turco University of Florence, Italy Pattern Recognition, 1998.
Lecture 4 Unsupervised Learning Clustering & Dimensionality Reduction
Advanced Topics in Data Mining Special focus: Social Networks.
1 An Empirical Study on Large-Scale Content-Based Image Retrieval Group Meeting Presented by Wyman
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
CS292 Computational Vision and Language Visual Features - Colour and Texture.
Ranking by Odds Ratio A Probability Model Approach let be a Boolean random variable: document d is relevant to query q otherwise Consider document d as.
Experimental Evaluation
HCC class lecture 14 comments John Canny 3/9/05. Administrivia.
Jeff Howbert Introduction to Machine Learning Winter Machine Learning Feature Creation and Selection.
Exercise Session 10 – Image Categorization
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Graph mining in bioinformatics Laur Tooming. Graphs in biology Graphs are often used in bioinformatics for describing processes in the cell Vertices are.
Census A survey to collect data on the entire population.   Data The facts and figures collected, analyzed, and summarized for presentation and.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
Author: M.E.J. Newman Presenter: Guoliang Liu Date:5/4/2012.
Liang Ge.  Introduction  Important Concepts in MCL Algorithm  MCL Algorithm  The Features of MCL Algorithm  Summary.
Scalable and Efficient Data Streaming Algorithms for Detecting Common Content in Internet Traffic Minho Sung Networking & Telecommunications Group College.
Glasgow 02/02/04 NN k networks for content-based image retrieval Daniel Heesch.
Particle Filters for Shape Correspondence Presenter: Jingting Zeng.
Universit at Dortmund, LS VIII
1 Motivation Web query is usually two or three words long. –Prone to ambiguity –Example “keyboard” –Input device of computer –Musical instruments How can.
Classifying Images with Visual/Textual Cues By Steven Kappes and Yan Cao.
80 million tiny images: a large dataset for non-parametric object and scene recognition CS 4763 Multimedia Systems Spring 2008.
IEEE Int'l Symposium on Signal Processing and its Applications 1 An Unsupervised Learning Approach to Content-Based Image Retrieval Yixin Chen & James.
The Dirichlet Labeling Process for Functional Data Analysis XuanLong Nguyen & Alan E. Gelfand Duke University Machine Learning Group Presented by Lu Ren.
Randomized Algorithms for Bayesian Hierarchical Clustering
Semantic Wordfication of Document Collections Presenter: Yingyu Wu.
Markov Cluster (MCL) algorithm Stijn van Dongen.
Andreas Papadopoulos - [DEXA 2015] Clustering Attributed Multi-graphs with Information Ranking 26th International.
INTERACTIVELY BROWSING LARGE IMAGE DATABASES Ronald Richter, Mathias Eitz and Marc Alexa.
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
Seminar on random walks on graphs Lecture No. 2 Mille Gandelsman,
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Complexity and Efficient Algorithms Group / Department of Computer Science Testing the Cluster Structure of Graphs Christian Sohler joint work with Artur.
Relation Strength-Aware Clustering of Heterogeneous Information Networks with Incomplete Attributes ∗ Source: VLDB.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Community structure in graphs Santo Fortunato. More links “inside” than “outside” Graphs are “sparse” “Communities”
1 Random Walks on the Click Graph Nick Craswell and Martin Szummer Microsoft Research Cambridge SIGIR 2007.
CS Machine Learning Instance Based Learning (Adapted from various sources)
Jianping Fan Department of Computer Science University of North Carolina at Charlotte Charlotte, NC Relevance Feedback for Image Retrieval.
Network Partition –Finding modules of the network. Graph Clustering –Partition graphs according to the connectivity. –Nodes within a cluster is highly.
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Non-parametric Methods for Clustering Continuous and Categorical Data Steven X. Wang Dept. of Math. and Stat. York University May 13, 2010.
Dynamic Network Analysis Case study of PageRank-based Rewiring Narjès Bellamine-BenSaoud Galen Wilkerson 2 nd Second Annual French Complex Systems Summer.
A Self-organizing Semantic Map for Information Retrieval Xia Lin, Dagobert Soergel, Gary Marchionini presented by Yi-Ting.
1 Dongheng Sun 04/26/2011 Learning with Matrix Factorizations By Nathan Srebro.
Hiroki Sayama NECSI Summer School 2008 Week 2: Complex Systems Modeling and Networks Network Models Hiroki Sayama
Large-Scale Content-Based Audio Retrieval from Text Queries
Efficient Image Classification on Vertically Decomposed Data
Nonparametric Semantic Segmentation
Machine Learning Feature Creation and Selection
Efficient Image Classification on Vertically Decomposed Data
Section 7.12: Similarity By: Ralucca Gera, NPS.
Junqi Zhang+ Xiangdong Zhou+ Wei Wang+ Baile Shi+ Jian Pei*
Machine Learning: UNIT-4 CHAPTER-1
Presentation transcript:

NN k Networks for browsing and clustering image collections Daniel Heesch Communications and Signal Processing Group Electrical and Electronic Engineering Imperial College London Robotics Research Group Seminar, Oxford

Structure Motivation Challenges The NN k Idea NN k Networks Clustering Conclusions and Outlook

Motivation 1k10k> 1Mio1k25k65k850k>1Mio

Motivation Providing access to large image collections based on visual image content not textual annotation Why are we so keen on content? - extraction can be automated - text annotation not exhaustive

Challenges Representation: How to capture meaningful content in terms of visual attributes? Scalability: How to alleviate the O(n) complexity of one-one matching? Polysemy: Which semantic facet in an image is the user interested in? And what is the best feature for representing it?

Challenges: Polysemy Tieu’s Convolution Feature Colour Structure Descriptor Global HSV Histogram

Semantic disambiguation Relevance Feedback: Users label retrieved images according to perceived relevance 1. Parametrise distance metric 2. find k-NN with default parameters 3. fine tune parameters upon relevance feedback

The NN k Idea Motivated by problems with traditional relevance feedback: - What is the weight set for the first retrieval? - Little scope for exploratory search 1. Parametrise distance metric 2. For every parameter set, record the closest image to some query image (subsequently referred to as focal image) This set of nearest neighbours defines the NN k of the query 3. Let users decide which parameter set is best by choosing among the retrieved images By considering all possible weighted combinations of features, we are more likely to capture the different semantic facets of the image

The NN k Idea Distance metric: where the k components of w are non-negative and sum to one. The weight space is a bounded and continuous set. To find the NN k of an image, we sample the weight space at a finite number of points.

The NN k Idea k = 2k = 3

The NN k Idea Sampling the weight space

NN k Networks Construct a directed graph by linking each image of a collection to all its NN k. Arc weight (x -> y): proportion of w for which y is NN k of x, i.e. relative size of supporting weight space

NN k Networks: Properties Distances: approx. 4 links between any two images (100,000 images), logarithmic increase with collection size Clustering coefficient (Watts & Strogatz, 2000): a vertex’s neighbours are likely to be neighbours themselves Precomputed, hence interaction is in real time Construction in O(n 2 ) time (14 hours for 50,000 images)

NN k Networks: Clustering NN k represent a range of semantic facets users may be interested in Once we know the semantic facet, we would like to present users with more images of the same kind Cluster the graph and display the cluster that contains the focal image and the chosen NN k Problem: a partitioning of the vertices allows the focal image to belong to only one cluster. We want a soft clustering A soft clustering can be achieved by partitioning not the vertex set but the edge set of the graph (a vertex can now associate with as many clusters as it has edges to other vertices) Compare with “Google Sets” and Zoubin Ghahramani’s talk two weeks ago

NN k Networks: Clustering MCL: Markov Clustering (van Dongen, 2000) Use normalised adjacency matrix as the transition matrix A of a random walk 1. Simulate a random walk: A t+1 = (A t ) I 2. Amplify large transition probabilities by raising each element of A to some power 3. Normalise resulting matrix and go to step 1 Matrix A converges to a sparse matrix. Interpreted as an adjacency matrix, it corresponds to a partitioning of the graph. MCL fast and suitable for directed graphs. Graphs with 100,000 vertices and 2,500,000 edges can be clustered in 1 hour (3GHz).

Dualising an NN k Network v w To speed up computation, consider only the r neighbours of vertex w with greatest weights. The dual of a graph with n vertices has at most rn vertices and r 2 n edges. x y z vw wx wy wz

NN k Networks: Clustering 32,000 images, r = 5, I = 20, 8 features

Cluster Validation With known classes, we may compare distribution of objects across clusters and classes Here, objects are assigned to several clusters and we don’t know the classes However, for Corel dataset we have rich image annotation available Assumption: good clusters tend to have a smaller joint vocabulary (description size) We can compare the observed description size with what we would expect if images were randomly assigned to clusters classes clusters

Cluster Validation What is the sampling distribution under the null hypothesis that images in a cluster are a random subset of the collection? If the number of terms is large and the clusters small, the description size is a sum of i.i.d. variates. By the Central Limit Theorem, we expect the sampling distribution to look Gaussian.

Cluster Validation For each cluster size, we generate a sampling distribution by recording the description size of 10,000 random subsets of that size The observed description size of a cluster can then be compared with that sampling distribution p-Value: probability that, given the null hypothesis is true, the description size is at least as low as the observed one

Conclusions NN k Networks as a representation of semantic relationships between images NN k Networks support fast interactive search but O(n 2 ) to construct Structure can be extracted by partitioning the dual of the graph. Even though each vertex’s neighbours form a diverse set, the subgraphs tend to be very uniform

Future Directions Scalability: Google provides access to > 1,3 Billion images Update: can we add an image without having to construct the whole network? Global features work well, greater discrimination could be achieved through domain-specific detectors (e.g. faces, background) Clustering: how does MCL fare in comparison with other techniques, e.g. spectral graph clustering?