Presentation is loading. Please wait.

Presentation is loading. Please wait.

Packing bag-of-features ICCV 2009 Herv´e J´egou Matthijs Douze Cordelia Schmid INRIA.

Similar presentations


Presentation on theme: "Packing bag-of-features ICCV 2009 Herv´e J´egou Matthijs Douze Cordelia Schmid INRIA."— Presentation transcript:

1 Packing bag-of-features ICCV 2009 Herv´e J´egou Matthijs Douze Cordelia Schmid INRIA

2 Introduction Proposed method Experiments Conclusion

3 Introduction Proposed method Experiments Conclusion

4 Bag-of-features Extracting local image descriptors Clustering of the descriptors & k-means quantizer(visual words) The histogram of visual word is weighted using the tf-idf weighting scheme of [12] & subsequently normalized with L2 norm Roducing a frequency vector fi of length k

5 TF–IDF weighting

6 TF–IDF weighting tf – 100 vocabularies in a document, ‘a’ 3 times – 0.03 (3/100) idf – 1,000 documents have ‘a’, total number of documents 10,000,000 – 9.21 ( ln(10,000,000 / 1,000) ) if-idf = 0.28( 0.03 * 9.21)

7 Binary BOF[12] discard the information about the exact number of occurrences of a given visual word in the image. Binary BOF vector components only indicates the presence or not of a particular visual word in the image. A sequential coding using 1 bit per component, ⌈ k/8 ⌉ bytes per image, the memory usage per image would be typically 10 kB per image [12] J. Sivic and A. Zisserman. Video Google: A text retrieval approach to object matching in videos. In ICCV, pages 1470–1477, 2003.

8 Binary BOF(Holidays dataset)

9 Inverted-file index(Sparsity) Documents – T 0 = "it is what it is" – T 1 = "what is it" – T 2 = "it is a banana" Index – "a": {2} – "banana": {2} – "is": {0, 1, 2} – "it": {0, 1, 2} – "what": {0, 1}

10 Binary BOF

11 Compressed inverted file Compression can close to the vector entropy Compared with a standard inverted file, about 4 times more images can be indexed using the same amount of memory This may compensate the decoding cost of the decompression algorithm [16] J. Zobel and A. Moffat. Inverted files for text search engines. ACM Computing Surveys, 38(2):6, 2006.

12 Introduction Proposed method Experiments Conclusion

13 MiniBOFs

14 Projection of a BOF Sparse projection matices – – d: dimension of the output descriptor – k: dimension of the input BOF For each matrix row, the number of non-zero components is, typically set nz = 8 for k = 1000, resulting in d = 125

15 Projection of a BOF The other matrices are defined by random permutations. – For k = 12 and d = 3, the random permutation (11, 2, 12, 8; 9, 4, 10, 1; 7, 5, 6, 3) Image i, m mini-BOFs –, ( )

16 Indexing structure Quantization – The miniBOF is quantized by associated with matrix,, where is the number of codebook entries of the indexing structure. – The set of k-means codebooks is learned off-line using a large number of miniBOF vectors, here extracted from the Flickr1M* dataset. The dictionary size associated with the minBOFs is not related to the one associated with the initial SIFT descriptors, hence we may choose. We typically set = 20000.

17 Indexing structure Binary signature generation – The miniBOF is projected using a random rotation matrix R, producing d components – Each bit of the vector is obtained by comparing the value projected by R to the median value of the elements having the same quantized index. The median values for all quantizing cells and all projection directions are learned off-line on our independent dataset

18 Quantizing cells [4] H. Jegou, M. Douze, and C. Schmid. Hamming embedding and weak geometric consistency for large scale image search. In ECCV, 2008.

19 Indexing structure miniBOF associated with image i is represented by the tuple total memory usage per image is bytes

20 Multi-probe strategy retrieving not only the inverted list associated with the quantized index, but the set of inverted lists associated with the closest t centroids of the quantizer codebook T times image hits

21 Fusion Query signature Database signature

22 Fusion – equal to 0 for images having no observed binary signatures – equal to if the database image i is the query image itself

23 Fusion

24 Introduction Proposed method Experiments Conclusion

25 Dataset Two annotated Dataset – INRIA Holidays dataset [4] – University of Ken-tucky recognition benchmark [9] Distractor dataset – one million images downloaded from Flickr, Flickr1M Learning parameters – Flickr1M ∗

26 Detail Descriptor extraction – Resize to a maximum of 786432 pixels – Performed a slight intensity normalization – SIFT Evaluation – Recall@N – mAP – Memory – Image hits Parameters # Using a value of nz between 8 and 12 provides the best accuracy for vocabulary sizes ranging from 1k to 20k.

27 mAP Mean average precision EX: – two images A&B – A has 4 duplicate images – B has 5 duplicate images – Retrieval rank A: 1, 2, 4, 7 – Retrieval rank B: 1, 3, 5 – Average precision A = (1/1+2/2+3/4+4/7)/4=0.83 – Average precision B = (1/1+2/3+3/5+0+0)/3=0.45 – mAP= (0.83+0.45)/2=0.64

28 Table 1(Holidays) # The number of bytes used per inverted list entry is 4 bytes for binary BOF & 5 bytes for BOF

29 Table 2(Kentucky)

30 Table 3(Holidays+Flickr1M)

31 Figure(Holidays+Flickr1M) # Our approach requires 160 MB for m = 8 and the query is performed in 132ms, to be compared, respectively, with 8 GB and 3s for BOF.

32 Sample

33 Introduction Proposed method Experiments Conclusion

34 This paper have introduced a way of packing BOFs:miniBOFs – An efficient indexing structure for rapid access and an expected distance criterion for the fusion of the scores – Reduces memory usage – Reduces the quantity of memory scanned (hits) – Reduces query time


Download ppt "Packing bag-of-features ICCV 2009 Herv´e J´egou Matthijs Douze Cordelia Schmid INRIA."

Similar presentations


Ads by Google