Students: Meera & Si Mentor: Afshin Dehghan WEEK 4: DEEP TRACKING.

Slides:



Advertisements
Similar presentations
Greedy Layer-Wise Training of Deep Networks
Advertisements

Scalable Learning in Computer Vision
Activity Recognition Aneeq Zia. Agenda What is activity recognition Typical methods used for action recognition “Evaluation of local spatio-temporal features.
Robust Object Tracking via Sparsity-based Collaborative Model
Deep Learning.
Effective Image Database Search via Dimensionality Reduction Anders Bjorholm Dahl and Henrik Aanæs IEEE Computer Society Conference on Computer Vision.
Graz University of Technology, AUSTRIA Institute for Computer Graphics and Vision Fast Visual Object Identification and Categorization Michael Grabner,
AN ANALYSIS OF SINGLE- LAYER NETWORKS IN UNSUPERVISED FEATURE LEARNING [1] Yani Chen 10/14/
Student: Kylie Gorman Mentor: Yang Zhang COLOR-ATTRIBUTES- RELATED IMAGE RETRIEVAL.
WEEK 6: DEEP TRACKING STUDENTS: SI CHEN & MEERA HAHN MENTOR: AFSHIN DEGHAN.
LLNL-PRES This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
Week 7: Deep Tracking Students: Si Chen & Meera Hahn Mentor: Afshin Deghan.
Deep Learning for Speech and Language Yoshua Bengio, U. Montreal NIPS’2009 Workshop on Deep Learning for Speech Recognition and Related Applications December.
Object Bank Presenter : Liu Changyu Advisor : Prof. Alex Hauptmann Interest : Multimedia Analysis April 4 th, 2013.
Project title : Automated Detection of Sign Language Patterns Faculty: Sudeep Sarkar, Barbara Loeding, Students: Sunita Nayak, Alan Yang Department of.
Video Tracking Using Learned Hierarchical Features
A shallow introduction to Deep Learning
University of Toronto Aug. 11, 2004 Learning the “Epitome” of a Video Sequence Information Processing Workshop 2004 Vincent Cheung Probabilistic and Statistical.
CVPR Workshop on RTV4HCI 7/2/2004, Washington D.C. Gesture Recognition Using 3D Appearance and Motion Features Guangqi Ye, Jason J. Corso, Gregory D. Hager.
Yang, Luyu.  Postal service for sorting mails by the postal code written on the envelop  Bank system for processing checks by reading the amount of.
Building high-level features using large-scale unsupervised learning Anh Nguyen, Bay-yuan Hsu CS290D – Data Mining (Spring 2014) University of California,
Mentor: Salman Khokhar Action Recognition in Crowds Week 7.
UCF REU: Weeks 1 & 2. Gradient Code Gradient Direction of the Gradient: Calculating theta.
Dr. Z. R. Ghassabi Spring 2015 Deep learning for Human action Recognition 1.
Deformable Part Model Presenter : Liu Changyu Advisor : Prof. Alex Hauptmann Interest : Multimedia Analysis April 11 st, 2013.
Particle Dynamics and Multi- Channel Feature Dictionaries for Robust Visual Tracking Srikrishna Karanam, Yang Li, Rich Radke Dept. of Electrical, Computer,
LeCun, Bengio, And Hinton doi: /nature14539
Week 10 Presentation Wesna LaLanne - REU Student Mahdi M. Kalayeh - Mentor.
Monitoring and Enhancing Visual Features (movement, color) as a Method for Predicting Brain Activity Level -in Terms of the Perception of Pain Sensation.
CT333/CT433 Image Processing and Computer Vision.
Week 10 Emily Hand UNR.
IEEE International Conference on Multimedia and Expo.
Learning video saliency from human gaze using candidate selection CVPR2013 Poster.
Convolutional Restricted Boltzmann Machines for Feature Learning Mohammad Norouzi Advisor: Dr. Greg Mori Simon Fraser University 27 Nov
Dynamic Background Learning through Deep Auto-encoder Networks Pei Xu 1, Mao Ye 1, Xue Li 2, Qihe Liu 1, Yi Yang 2 and Jian Ding 3 1.University of Electronic.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Hierarchical Motion Evolution for Action Recognition Authors: Hongsong Wang, Wei Wang, Liang Wang Center for Research on Intelligent Perception and Computing,
Week 3 Emily Hand UNR. Online Multiple Instance Learning The goal of MIL is to classify unseen bags, instances, by using the labeled bags as training.
Xintao Wu University of Arkansas Introduction to Deep Learning 1.
Yann LeCun Other Methods and Applications of Deep Learning Yann Le Cun The Courant Institute of Mathematical Sciences New York University
Learning Multi-Domain Convolutional Neural Networks for Visual Tracking arXiv : [cs.CV] v1 2015, v Hyeonseob Nam, Bohyung Han Dept.
Unsupervised Learning of Video Representations using LSTMs
Machine Learning for Data Certification at CMS
Deep Learning Amin Sobhani.
Automatic Lung Cancer Diagnosis from CT Scans (Week 2)
A practical guide to learning Autoencoders
Saliency-guided Video Classification via Adaptively weighted learning
Week III: Deep Tracking
Performance of Computer Vision
Article Review Todd Hricik.
Automatic Lung Cancer Diagnosis from CT Scans (Week 4)
Mentor: Afshin Dehghan
Comparison Between Deep Learning Packages
Unsupervised Learning and Autoencoders
Action Recognition in Temporally Untrimmed Videos
Convolutional Neural Networks for Visual Tracking
Jia-Bin Huang Virginia Tech ECE 6554 Advanced Computer Vision
Video understanding using part based object detection models
HCI/ComS 575X: Computational Perception
Deep learning Introduction Classes of Deep Learning Networks
KFC: Keypoints, Features and Correspondences
Hand Gesture Recognition Using Hidden Markov Models
Age and Gender Classification using Convolutional Neural Networks
Object Classification through Deconvolutional Neural Networks
AHED Automatic Human Emotion Detection
Multiple Organ detection in CT Volumes Using Random Forests - Week 5
Report 4 Brandon Silva.
Week 3 Volodymyr Bobyr.
Counting in High-Density Crowd Videos
Presentation transcript:

Students: Meera & Si Mentor: Afshin Dehghan WEEK 4: DEEP TRACKING

CURRENT PROGRESS

HANDCRAFTED FEATURES VS AUTOENCODERS 1 Online Object Tracking: A Benchmark. Yi Wu, Jongwoo Lim, and Ming-Hsuan Yang, “Online Object Tracking: A Benchmark,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Downloaded 10 videos from Online Object Tracking: A Benchmark 1, and cut them to 115 frames Compared autoencoder results with HOG and Color Histogram HOG performed the worst overall Autoencoders performed the best in all but 2 videos

VISUALIZING THE FILTERS Currently running code with 3 layers Could not visualize the 2 nd and 3 rd layers Could visualize the 1 st layer 2 nd and 3 rd layer representations are different from the 1st The current visualization function does not apply to them

APPLYING GAUSSIAN CONFIDENCE BASED ON MOTION VECTOR Performance for the sequences was improved Change in confidence values We observed the effect of a Gaussian motion model

NEXT STEPS

Training 2 Finish downloading 1 million images of same size Pre-train network with the images Fine tune the network Visualizing Layers Currently we can only visualize the 1 st layer of filters Do more research and implement a method to visualize 2 nd and 3 rd layers NEXT STEPS 2 Lamblin, Pascal and Yoshua Bengio. Important Gains from Supervised Fine-Tuning of Deep Architectures on Large Labeled Sets. NIPS 2010 Deep Learning and Unsupervised Feature Learning Workshop.

Variations Patch size Greyscale images Further Reading Fine tuning networks Visualizing filters of higher layers Learning motion: provide temporal data to network so it can learn the vector NEXT STEPS