REU Report Meeting Week 9

Slides:



Advertisements
Similar presentations
Distinctive Image Features from Scale-Invariant Keypoints David Lowe.
Advertisements

Recognizing Human Actions by Attributes CVPR2011 Jingen Liu, Benjamin Kuipers, Silvio Savarese Dept. of Electrical Engineering and Computer Science University.
DIMENSIONALITY REDUCTION: FEATURE EXTRACTION & FEATURE SELECTION Principle Component Analysis.
Large-scale, real-world facial recognition in movie trailers Alan Wright Presentation 8.
Proactive Learning: Cost- Sensitive Active Learning with Multiple Imperfect Oracles Pinar Donmez and Jaime Carbonell Pinar Donmez and Jaime Carbonell Language.
Bring Order to Your Photos: Event-Driven Classification of Flickr Images Based on Social Knowledge Date: 2011/11/21 Source: Claudiu S. Firan (CIKM’10)
Watching Unlabeled Video Helps Learn New Human Actions from Very Few Labeled Snapshots Chao-Yeh Chen and Kristen Grauman University of Texas at Austin.
Retrieving Actions in Group Contexts Tian Lan, Yang Wang, Greg Mori, Stephen Robinovitch Simon Fraser University Sept. 11, 2010.
Beyond Actions: Discriminative Models for Contextual Group Activities Tian Lan School of Computing Science Simon Fraser University August 12, 2010 M.Sc.
ACM SAC’06, DM Track Dijon, France “The Impact of Sample Reduction on PCA-based Feature Extraction for Supervised Learning” by M. Pechenizkiy,
Machine Learning: Final Presentation James Dalphond James McCauley Andrew Wilkinson Phil Kovac Data Set: Yeast GOLD TEAM.
Evaluation of Utility of LSA for Word Sense Discrimination Esther Levin, Mehrbod Sharifi, Jerry Ball
Large-Scale Content-Based Image Retrieval Project Presentation CMPT 880: Large Scale Multimedia Systems and Cloud Computing Under supervision of Dr. Mohamed.
Classifiers, Part 3 Week 1, Video 5 Classification  There is something you want to predict (“the label”)  The thing you want to predict is categorical.
Overcoming Dataset Bias: An Unsupervised Domain Adaptation Approach Boqing Gong University of Southern California Joint work with Fei Sha and Kristen Grauman.
Internet-scale Imagery for Graphics and Vision James Hays cs195g Computational Photography Brown University, Spring 2010.
1 Comparison of Principal Component Analysis and Random Projection in Text Mining Steve Vincent April 29, 2004 INFS 795 Dr. Domeniconi.
Lecture 31: Modern recognition CS4670 / 5670: Computer Vision Noah Snavely.
Classification Course web page: vision.cis.udel.edu/~cv May 12, 2003  Lecture 33.
Chao-Yeh Chen and Kristen Grauman University of Texas at Austin Efficient Activity Detection with Max- Subgraph Search.
Geodesic Flow Kernel for Unsupervised Domain Adaptation Boqing Gong University of Southern California Joint work with Yuan Shi, Fei Sha, and Kristen Grauman.
Intelligent Database Systems Lab Presenter: Wu, Jhen-Wei Authors: Fabian Bürger, Josef Pauli ICPRAM. Representation Optimization with Feature Selection.
A feature-based kernel for object classification P. Moreels - J-Y Bouguet Intel.
WEEK4 RESEARCH Amari Lewis Aidean Sharghi. PREPARING THE DATASET  Cars – 83 samples  3 images for each sample when x=0  7 images for each sample when.
Category Category Category Category Category
Does one size really fit all? Evaluating classifiers in a Bag-of-Visual-Words classification Christian Hentschel, Harald Sack Hasso Plattner Institute.
1 Bilinear Classifiers for Visual Recognition Computational Vision Lab. University of California Irvine To be presented in NIPS 2009 Hamed Pirsiavash Deva.
P.Demestichas (1), S. Vassaki(2,3), A.Georgakopoulos(2,3)
PREDICT 422: Practical Machine Learning
COMP 1942 PCA TA: Harry Chan COMP1942.
School of Computer Science & Engineering
Article Review Todd Hricik.
An Empirical Study of Learning to Rank for Entity Search
Statistical Techniques
TeamMember1 TeamMember2 Machine Learning Project 2016/2017-2
Mixture of SVMs for Face Class Modeling
Project 2 k-NN 2018/11/10.
Outline Multilinear Analysis
Video Summarization via Determinantal Point Processes (DPP)
Cheng-Ming Huang, Wen-Hung Liao Department of Computer Science
Machine Learning for Online Query Relaxation
Finding Clusters within a Class to Improve Classification Accuracy
Location Recommendation — for Out-of-Town Users in Location-Based Social Network Yina Meng.
Bird-species Recognition Using Convolutional Neural Network
CS 1674: Intro to Computer Vision Scene Recognition
Lesson 3. 1 How do we interpret and represent functions. Standards F
CSSE 371, Fall, 2013 “What to do for the client meeting” Week 4
Open-Category Classification by Adversarial Sample Generation
Levels of Organisation. YouTube Video – Levels of Organisation
المشرف د.يــــاســـــــــر فـــــــؤاد By: ahmed badrealldeen
Effective learning strategies
Intro to Machine Learning
Cross-validation Brenda Thomson/ Peter Fox Data Analytics
Housam Babiker, Randy Goebel and Irene Cheng
Evaluating Classifiers
CLS, Rapid Schema Consistent Learning, and Similarity-weighted Interleaved learning Psychology 209 Feb 26, 2019.
Assignment 1: Classification by K Nearest Neighbors (KNN) technique
Helen: Maliciously Secure Coopetitive Learning for Linear Models
MAS 622J Course Project Classification of Affective States - GP Semi-Supervised Learning, SVM and kNN Hyungil Ahn
Xiao-Yu Zhang, Shupeng Wang, Xiaochun Yun
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
Lecture 16. Classification (II): Practical Considerations
Extracting Why Text Segment from Web Based on Grammar-gram
Jiahe Li
CRCV REU 2019 Week 5.
Self-Supervised Cross-View Action Synthesis
Self-Supervised Cross-View Action Synthesis
Machine Learning: Lecture 5
Making Inferences.
Truman Action Recognition Status update
Presentation transcript:

REU Report Meeting Week 9 Patrick Hynes

Outline – One Shot Recap This Week’s Results Next Steps YouTube Dataset Caltech Next Steps

YouTube Dataset 11 Categories, 100 Videos each

Methods One-Shot Learning of Classes Leave-One-Out Cross Validation

5 Train , 6 Test Dimension Reduction 1 Shot 3 Shot 6 Shot 15 Shot PCA 20.04 20.41 21.71 23.66 Fast SVD 27.42 29.04 33.05 36.21 29.26 --- 34.59 32.83 38.16 27.60 31.18 32.03 37.58 33.90 38.52

“Videos in wild” Results Overall 71.5% Leave-One-Out Cross Validation

Baseline How well could we do, training on everything? Training Classes Testing Classes Dimension Reduction 1 Shot 3 Shot 6 Shot 15 Shot 11 PCA 85.40 --- 92.81 Fast SVD 81.68 89.04 91.69 93.24

Can we compare more directly Train and Test Using all Classes, but separate examples pulled randomly Train Examples Test Examples Dimension Reduction 1 Shot 15 Shot 75 25 PCA 59.57 68.15 SVDS 63.64 72.68 Fast SVD 69.85 79.80 50 66.80 71.26 61.68 64.07

Cross Validation Results Testing examples pulled in order Testing examples pulled randomly 1 Shot Mean 15 Shot Mean 90-10 Fast SVD 54.03 63.38 1 Shot Mean 15 Shot Mean 90-10 Fast SVD 64.57 75.91

Person by Person Cross Validation “96-4” Neighbors Optimization 3 sets, 1 train, 10 test 1 ‘shot’ 50.85 15 ‘shot 59.70 25 ‘shot’ 60.14 96 ‘shot’ 61.28

Conclusions Method is heavily disturbed by these different videos Results show some possible inconsistencies that may need to be explored

Caltech Experiments Try different dimension reductions strategy Anthony will have additional results on his Caltech experiments

Recall Trained on Caltech 256 One Shot on Caltech 101 Neighbors Our Method Fei-Fei, et all Grauman and Darrell 1 9.73% ---- 17.5% 3 11.00% 10.4% 28% 6 14.73% 13.9%

FastSVD vs. PCA Training on 256, Testing 101 FastSVD doesn’t give a similar benefit in this instance Neighbors PCA FastSVD 1 9.73% 5.40% 3 11.00% 5.97% 6 14.73% 7.36%

Next Steps Can we do better at recognizing an action in a very different video? Reevaluate features How can we improve on our one shot results and use that knowledge to help answer the first question? Run additional one-shot experiments Continue one-shot experiments with the Caltech datasets.