Presentation is loading. Please wait.

Presentation is loading. Please wait.

Glasgow 02/02/04 NN k networks for content-based image retrieval Daniel Heesch.

Similar presentations


Presentation on theme: "Glasgow 02/02/04 NN k networks for content-based image retrieval Daniel Heesch."— Presentation transcript:

1 Glasgow 02/02/04 NN k networks for content-based image retrieval Daniel Heesch

2 Overview CBIR – a problem of image understanding? Approaches to feature weighting The NN k technique for retrieval Getting connected: NN k networks for browsing Future work

3 Challenges: Semantic gap What is the relationship between low-level features and image meaning?

4 Challenges: Image polysemy one image - multiple meanings

5 Feature Weights

6 Approaches to feature weighting (1) Post retrieval relevance feedback effectiveness relies on a good first retrieval result useful for fine-tuning weights

7 Approaches to feature weighting (2) SVM Metaclassifier ( Yavlinsky et al., ICASSP 2004 ) Given a set of queries and ground truth, for each query: sample at random m positive and negative images and for each build a score vector consisting of the feature- specific similarities between that image and the query Use an SVM to obtain a hyperplane that separates positive and negative score vectors with least error For an unseen image, compute similarity to the query as the distance of its score vector to the hyperplane.

8 Approaches to feature weighting (2) The distance of a vector to a hyperplane is a weighted sum of the vector’s components, which is just the aggregation formula shown to you previously. The hyperplane represents a set of feature weights that maximise the expected mean average precision

9 Approaches to feature weighting (2) ~ 300 score vectors needed to establish near-optimal weights for a subset of the Corel collection on average (6192 images) No query-specific weight optimization

10 Approaches to feature weighting (3) Query-specific optimization (Aggarwal et al., IEEE Trans. Multimedia 2002) Modify query representation along each feature axis and regenerate modified query, ask user whether new query image is still relevant Interesting idea but limited applicability in practice

11 NN k retrieval

12 The idea of NN k – a two-step approach 1.Retrieve with all possible weight sets -> returns a set of images (NN k ) each associated with a particular weight 2.Retrieve with the weights associated with the relevant images the user selects

13 The first retrieval step: finding the NN k For each feature combination w, determine nearest neighbour of the query Record for each nearest neighbour the proportion of w for which it came top as well as the average of these w NN for nearest neighbour, k for the dimensionality of the weight space (= length of the weight vector w)

14 F2 F1 F3 The first retrieval step: finding the NN k 1 1 1

15

16 With fixed number of grid points, time complexity is exponential in the number of features (k) Useful theorem: if for any two weight sets w 1 and w 2 that differ only in two components the top ranked image is the same, then this image will be top ranked for all linear combinations of w 1 and w 2

17 The first retrieval step: finding the NN k

18 Visualization of NN k

19 Retrieve with each weight set in turn Merge ranked lists The second retrieval step:

20 Comparison of NN k with two post-retrieval methods for weight-update 1. Our own: minimize 2. Rui’s method (Rui et al., 2002) Performance evaluation

21 Corel Gallery 380,000 Package Given a subset of images, treat each image as a query in turn and retrieve from the rest For RF: retrieve with equal weight sets, gather relevance data and retrieve with new weight set For NN k : determine NN k, gather relevance data and retrieve with new weight sets Determine MAP after second retrieval

22 Performance evaluation

23 NN k networks

24 Network Construction Vertices represent images An arc is established between two images X and Y, iff there exist at least one instantiation of the weight vector w, for which Y is the nearest neighbour of X Record for each nearest neighbour the proportion of w, for which it came top -> edge weight, measure of similarity Storage: for each image, its nearest neighbours and their frequencies

25 Rationale exposure of semantic richness user decides which image meaning is the correct one network precomputed -> interactive supports search without query formulation

26 Graph topology: small world properties small average distance between any two vertices (three nodes for 32000 images) high clustering coefficient: an image‘s neighbours are likely to be neighbours themselves CorelSketchesVideo k547 Vertices6,19223832,318 Edges150,7761,8221,253,076 C(G)0.0470.1340.14 C rand (G)0.0040.030.0012 Dist3.223.293.33 Dist rand 2.732.682.83

27 Graph topology: scale-freeness Degree distribution follows power-law

28 Image Browsing Initial display: retrieval result using search-by-example OR cluster display using Markov-Chain Clustering (MCL) technique (van Dongen, 2000) Clicking on an image displays all adjacent vertices in the network Distance inversely proportional to edge weight

29

30

31 Evaluation of NN k networks: TRECVID2003 search collection: 32000 keyframes from news videos 24 topics: example images + text Four system variants: Search + Relevance Feedback + Browsing Search + Relevance Feedback Search + Browsing Browsing

32

33 Future work

34 User interaction What can we infer from the history of images in the user‘s search path? The location of the target? Changing information needs?

35 Network construction and analysis Hierarchical sampling of points in weight space Incremental update of network while preserving small-world properties Optimal network structures

36 Thanks


Download ppt "Glasgow 02/02/04 NN k networks for content-based image retrieval Daniel Heesch."

Similar presentations


Ads by Google