AbstractAbstract Adjusting Active Basis Model by Regularized Logistic Regression Ruixun Zhang Peking University, Department of Statistics and Probability.

Slides:



Advertisements
Similar presentations
Adjusting Active Basis Model by Regularized Logistic Regression
Advertisements

Detecting Faces in Images: A Survey
Rapid Object Detection using a Boosted Cascade of Simple Features Paul Viola, Michael Jones Conference on Computer Vision and Pattern Recognition 2001.
Pattern Recognition and Machine Learning
SVM—Support Vector Machines
Optimization Tutorial
An Overview of Machine Learning
Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem
Lecture 31: Modern object recognition
Steerable Part Models Hamed Pirsiavash and Deva Ramanan
Computer vision: models, learning and inference
Cos 429: Face Detection (Part 2) Viola-Jones and AdaBoost Guest Instructor: Andras Ferencz (Your Regular Instructor: Fei-Fei Li) Thanks to Fei-Fei Li,
Fei Xing1, Ping Guo1,2 and Michael R. Lyu2
Data Visualization STAT 890, STAT 442, CM 462
AdaBoost & Genetic algorithms: application to pedestrian detection Yotam Abramson Ecole des Mines de Paris 9/12/05 Korea-France SafeMove Workshop.
Discriminative and generative methods for bags of features
Robust Moving Object Detection & Categorization using self- improving classifiers Omar Javed, Saad Ali & Mubarak Shah.
Beyond bags of features: Part-based models Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba.
Beyond bags of features: Adding spatial information Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba.
Logistic Regression Rong Jin. Logistic Regression Model  In Gaussian generative model:  Generalize the ratio to a linear model Parameters: w and c.
A Brief Introduction to Adaboost
Boosting Demo Object Detection Using the Statistics of Parts and Robust Real-time Object Detection Peter Barnum March 6, 2006 Advanced Perception.
1 How to be a Bayesian without believing Yoav Freund Joint work with Rob Schapire and Yishay Mansour.
A Robust Real Time Face Detection. Outline  AdaBoost – Learning Algorithm  Face Detection in real life  Using AdaBoost for Face Detection  Improvements.
The Chicken Project Dimension Reduction-Based Penalized logistic Regression for cancer classification Using Microarray Data By L. Shen and E.C. Tan Name.
Learning Programs Danielle and Joseph Bennett (and Lorelei) 4 December 2007.
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
CIVS, Statistics Dept. UCLA Deformable Template as Active Basis Zhangzhang Si UCLA Department of Statistics Ying Nian Wu, Zhangzhang Si, Chuck.
Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem
Learning Based Hierarchical Vessel Segmentation
Data mining and machine learning A brief introduction.
Learning Visual Bits with Direct Feature Selection Joel Jurik 1 and Rahul Sukthankar 2,3 1 University of Central Florida 2 Intel Research Pittsburgh 3.
EADS DS / SDC LTIS Page 1 7 th CNES/DLR Workshop on Information Extraction and Scene Understanding for Meter Resolution Image – 29/03/07 - Oberpfaffenhofen.
Computer Vision CS 776 Spring 2014 Recognition Machine Learning Prof. Alex Berg.
Recognition using Boosting Modified from various sources including
2 2  Background  Vision in Human Brain  Efficient Coding Theory  Motivation  Natural Pictures  Methodology  Statistical Characteristics  Models.
Unsupervised Learning of Compositional Sparse Code for Natural Image Representation Ying Nian Wu UCLA Department of Statistics October 5, 2012, MURI Meeting.
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
A Statistically Selected Part-Based Probabilistic Model for Object Recognition Zhipeng Zhao, Ahmed Elgammal Department of Computer Science, Rutgers, The.
Window-based models for generic object detection Mei-Chen Yeh 04/24/2012.
Lecture 31: Modern recognition CS4670 / 5670: Computer Vision Noah Snavely.
Image Categorization Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 03/15/11.
Classifiers Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 04/09/15.
Transfer Learning for Image Classification Group No.: 15 Group member : Feng Cai Sauptik Dhar Sauptik.
ECE738 Advanced Image Processing Face Detection IEEE Trans. PAMI, July 1997.
An Ensemble of Three Classifiers for KDD Cup 2009: Expanded Linear Model, Heterogeneous Boosting, and Selective Naive Bayes Members: Hung-Yi Lo, Kai-Wei.
Object Category Detection: Sliding Windows Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 03/18/10.
Face Detection Ying Wu Electrical and Computer Engineering Northwestern University, Evanston, IL
Christopher M. Bishop, Pattern Recognition and Machine Learning.
Classification Derek Hoiem CS 598, Spring 2009 Jan 27, 2009.
Latent Variable / Hierarchical Models in Computational Neural Science Ying Nian Wu UCLA Department of Statistics March 30, 2011.
Lecture 6: Classification – Boosting and SVMs CAP 5415 Fall 2006.
Methods for classification and image representation
Cross Validation of SVMs for Acoustic Feature Classification using Entire Regularization Path Tianyu Tom Wang T. Hastie et al WS04.
Christopher M. Bishop Object Recognition: A Statistical Learning Perspective Microsoft Research, Cambridge Sicily, 2003.
Notes on HW 1 grading I gave full credit as long as you gave a description, confusion matrix, and working code Many people’s descriptions were quite short.
Convolutional Restricted Boltzmann Machines for Feature Learning Mohammad Norouzi Advisor: Dr. Greg Mori Simon Fraser University 27 Nov
Statistical Template-Based Object Detection A Statistical Method for 3D Object Detection Applied to Faces and Cars Henry Schneiderman and Takeo Kanade.
Machine Learning: A Brief Introduction Fu Chang Institute of Information Science Academia Sinica ext. 1819
Perceptual and Sensory Augmented Computing Machine Learning, Summer’09 Machine Learning – Lecture 7 Model Combination & Boosting Bastian Leibe.
Statistical Modeling and Learning in Vision --- cortex-like generative models Ying Nian Wu UCLA Department of Statistics JSM, August 2010.
General Information Course Id: COSC6342 Machine Learning Time: TU/TH 1-2:30p Instructor: Christoph F. Eick Classroom:AH301
Face Detection 蔡宇軒.
PREDICT 422: Practical Machine Learning
Boosting and Additive Trees
Special Topics in Data Mining Applications Focus on: Text Mining
Object detection as supervised classification
Finding Clusters within a Class to Improve Classification Accuracy
Cos 429: Face Detection (Part 2) Viola-Jones and AdaBoost Guest Instructor: Andras Ferencz (Your Regular Instructor: Fei-Fei Li) Thanks to Fei-Fei.
Derek Hoiem CS 598, Spring 2009 Jan 27, 2009
Presentation transcript:

AbstractAbstract Adjusting Active Basis Model by Regularized Logistic Regression Ruixun Zhang Peking University, Department of Statistics and Probability References Wu, Y. N., Si, Z., Gong, H. and Zhu, S.-C. (2009). Learning Active Basis Model for Object Detection and Recognition. International Journal of Computer Vision. R.-E. Fan, K.-W. Chang, C.-J. Hsieh, X.-R. Wang, and C.-J. Lin. (2008). LIBLINEAR: A Library for Large Linear Classification. Journal of Machine Learning Research. Lin, C. J., Weng, R.C., Keerthi, S.S. (2008). Trust Region Newton Method for Large-Scale Logistic Regression. Journal of Machine Learning Research. Vapnik, V. N. (1995). The Nature of Statistical Learning Theory. Springer. Joachims, T. (1999). Making large-Scale SVM Learning Practical. Advances in Kernel Methods - Support Vector Learning, MIT-Press. Freund, Y. and Schapire, R. E. (1997). A Decision-Theoretic Generalization of On- Line Learning and an Application to Boosting. Journal of Computer and System Sciences. Viola, P. and Jones, M. J. (2004). Robust real-time face detection. International Journal of Computer Vision. Rosset, S., Zhu, J., Hastie, T. (2004). Boosting as a Regularized Path to a Maximum Margin Classifier. Journal of Machine Learning Research. Zhu, J. and Hastie, T. (2005). Kernel Logistic Regression and the Import Vector Machine. Journal of Computational and Graphical Statistics. Hastie, T., Tibshirani, R. and Friedman, J. (2001) Elements of Statistical Learning; Data Mining, Inference, and Prediction. New York: Springer. Bishop, C. (2006). Pattern Recognition and Machine Learning. New York: Springer. L. Fei-Fei, R. Fergus and P. Perona. (2004). Learning generative visual models from few training examples: an incremental Bayesian approach tested on 101 object categories. IEEE. CVPR, Workshop on Generative-Model Based Vision. Friedman, J., Hastie, T. and Tibshirani, R. (2000). Additive logistic regression: A statistical view of boosting (with discussion). Ann. Statist. MethodsMethods IntroductionIntroduction ResultsResults DiscussionDiscussion AcknowledgementsAcknowledgements Active basis model is a generative model seeking a common wavelet sparse coding of images from the same object category, where the images share the same set of selected wavelet elements, which are allowed to perturb their locations and orientations to account for shape deformations. This work applies discriminative methods to adjust λ’s of selected basis elements, including logistic regression, SVM and AdaBoost. Results on supervised learning show that discriminative post-processing on active basis model improves its classification performance in terms of testing AUC. Among the three methods the L2-regularized logistic regression is the most natural one and performs the best.  Active Basis – Generative Model a common template for a set of images: perturbed to match each image. where Basis perturbation A set of images share an active basis  Unsupervised Learning Unknown categories: Unknown locations and scales:  Discriminative Adjustment Adjust λ’s of the template by: L2-regularized logistic regression: SVM and AdaBoost: Over-fitting without regularization. Hidden variables Model: Loss function:  Classification Experiment Template size = 80. Tuning parameter = Head_shoulder data: training negatives 160, testing negatives 471. Guitar data: training negatives 160, testing negatives 855.  Tuning Parameter Small tuning parameters imply high regularization guarantee high performances  Future Work Extend to unsupervised learning – adjust mixture model Generative learning by active basis Discriminative adjustment on feature weights Thanks to my mentor Prof. Ying Nian Wu and PhD fellow Zhangzhang Si, for their instructions to my project, as well as to my self developing. I have learned a lot from the project. It is a fantastic summer for me. Also thanks to Dr. Chih-Jen Lin for his liblinear software package and his detailed suggestions about how to adjust the software for our experiment.