Presentation is loading. Please wait.

Presentation is loading. Please wait.

Object Bank Presenter : Liu Changyu Advisor : Prof. Alex Hauptmann Interest : Multimedia Analysis April 4 th, 2013.

Similar presentations


Presentation on theme: "Object Bank Presenter : Liu Changyu Advisor : Prof. Alex Hauptmann Interest : Multimedia Analysis April 4 th, 2013."— Presentation transcript:

1 Object Bank Presenter : Liu Changyu Advisor : Prof. Alex Hauptmann Interest : Multimedia Analysis April 4 th, 2013

2 CMU - Language Technologies Institute 2 Contents Introduction Model Algorithm Experiment Conclusion

3 CMU - Language Technologies Institute 3 1. Research Question 1) Understanding the meanings and contents of images remains one of the most challenging problems in machine intelligence and statistical learning. 2) Also present low-level image features are enough for a variety of visual recognition tasks, but still not enough, especial for the visual tasks which carry semantic meanings. So efficient high level image features are often needed. Introduction

4 CMU - Language Technologies Institute 4 2. What’s Object Bank? Object bank representation is a novel image representation for high-level visual tasks, which encodes semantic and spatial information of the objects within an image. In object bank, an image is represented as a collection of scale-invariant response maps of a large number of pre-trained generic object detectors. Introduction

5 CMU - Language Technologies Institute 5 3. Why to use it? Fig.1 illustrates the gradient-based GIST features and texture-based Spatial Pyramid representation of two different scenes (foresty mountain vs. street). But such schemes often fail to offer sufficient discriminative power, as one can see from the very similar image statistics in the examples in this figure. Introduction Fig. 1 : (Best viewed in colors and magnification.) Comparison of OB) representation with GIST and SIFT-SPM of mountain vs. city street..

6 CMU - Language Technologies Institute 6 4. What is it used for? The main goal we want to use object bank is: 1) Optimize object bank detection code 2) Extend object banks to incorporate more objects. Introduction

7 CMU - Language Technologies Institute 7 Contents Introduction Model Algorithm Experiment Conclusion

8 CMU - Language Technologies Institute 8 Fig. 2 Object Bank Model Model ---Object Bank A large number of object detectors are first applied to an input image at multiple scales. For each object at each scale, a three-level spatial pyramid representation of the resulting object filter map is used; the maximum response for each object in each grid is then computed, resulting in a No:Objects length feature vector for each grid. A concatenation of features in all grids leads to an OB descriptor for the image.

9 CMU - Language Technologies Institute 9 Contents Introduction Model Algorithm Experiment Conclusion

10 CMU - Language Technologies Institute 10 1)Let represent the design built on the J-dimensional object bank representation of N images; 2) Let denote the binary classification labels of N samples. 3) This leads to the following learning problem: (1) Where is some non-negative, convex loss is a regularizer that avoids overfitting. Algorithm According Paper [1]Object Bank Algorithm is as follows:

11 CMU - Language Technologies Institute 11 Contents Introduction Model Algorithm Experiment Conclusion

12 CMU - Language Technologies Institute 12 We want to extend the original Object Bank approach, and do some related experiments as follow steps: 1) List and number the needed objects in our experiment, as: Object Names: 10200-clock 10302-goggles 10306-spectacles 10477-knife 10572-key 10577-keyboard 10638-desktop computer 10658-computer 1074-dog 10790-printer 10887-faucet ……………………… Experiment

13 CMU - Language Technologies Institute 13 2) Download the related bounding box from image-net. 3) Resizing the original image: The image is resized using the following process. First get the image dimensions(i.e. (a,b)). The ratio for scaling is calculated, using the following: Ratio=400/min(a,b); Therefore, the smaller axe of the image is converted to 400 pixels. This example illustrates that: Experiment Fig. 3 Resizing Step

14 CMU - Language Technologies Institute 14 4) Getting HOG features at different scales: After this rescaling, HOG features are obtained using different scales of the image. Although, they obtain HOG features for more scales, they only use six of these scales. These are the ratios used for resizing the image(previously resized in the previous step) Ratios: 1(image obtained from the previous step) 0.707 0.5 0.3535 0.25 0.17677 0.128 Experiment

15 CMU - Language Technologies Institute 15 After resizing the images, then calculate the HOG features for every image. These HOG features are used to obtain the response for every object. Example of HOG features are calculated for one image: Experiment Fig. 4 Example of HOG feature

16 CMU - Language Technologies Institute 16 5) Getting the response for the object): After getting the HOG features, we apply a object specific filter to these features. Each root filter, has two different components. Each of these components works on a different scale. As a result, we have 12 different detection scales because we obtained had 6 different scales from the previous steps, and every filter works at 2 different scales. Consequently, 6*2=12. These filter responses, are stored in a matrixes following the same distribution as the HOG feature in the image d of the previous figure. Then, for the HOG obtained from each ratio we have two different filter responses. Namely, we have 12 HOG responses. Experiment

17 CMU - Language Technologies Institute 17 6) Getting the spatial pyramids: We have 3 different spatial pyramid levels: These three spatial pyramid levels are applied to the 12 different responses to one object. In order to select the value for each box, they select the maximum response of the filter for every box. For instance, for the second level, they split the filter response using a grid of 2x2; They pick the maximum response inside every box. As you can see, we have 21( 1 + 2*2 + 4*4) values for every one of the 12 filter responses for one object. Resulting in a total of 12*21= 252 dimensions vector for every object. Experiment

18 CMU - Language Technologies Institute 18 7) Getting the feature vector for one object): Now, I am going to describe the distribution of the feature vector for one object. We have a vector of 252 dimensions. We start for the diferent scales(Remember that our original image, is the one obtained in the first step) Experiment

19 CMU - Language Technologies Institute 19 Each one of the chunk for every scale is divided in two pieces, because the used root filter in object bank has two different components that work at different scales. So these 42 dimensions for every scale are splited in two pieces of 21 dimensions. Experiment

20 CMU - Language Technologies Institute 20 Finally, This is the distribution of the 21 dimensions. Experiment

21 CMU - Language Technologies Institute 21 Contents Introduction Model Algorithm Experiment Conclusion

22 CMU - Language Technologies Institute 22 1)It is a feasible method to use. The author used several experiments to demonstrate that Object Bank representation that carries rich semantic level image information is more powerful on scene classification tasks than some other popular methods. 2)We could use and extend this approach according the real situation to do the remain experiment in the near future. Conclusion

23 CMU - Language Technologies Institute 23 Reference [1] Level Image Representation for Scene Classification and Semantic Feature Sparsification. Proceedings of the Neural Information Processing Systems (NIPS), 2010. [2] Li-Jia Li, Hao Su, Yongwhan Lim and Li Fei-Fei. Objects as Attributes for Scene Classification. Proceedings of the 12th European Conference of Computer Vision (ECCV), 1st International Workshop on Parts and Attributes, 2010. [3] Sreemanananth Sadanand and Jason J. Corso. Action Bank: A High-Level Representation of Activity in Video. CVPR, 2012. [4]Pedro Felzenszwalb, et. al. A Discriminatively Trained, Multiscale, Deformable Part Model. CVPR, 2008.

24 CMU - Language Technologies Institute 24 Thank you!


Download ppt "Object Bank Presenter : Liu Changyu Advisor : Prof. Alex Hauptmann Interest : Multimedia Analysis April 4 th, 2013."

Similar presentations


Ads by Google