Graph-Based Semi-Supervised Learning with a Generative Model Speaker: Jingrui He Advisor: Jaime Carbonell Machine Learning Department 04-10-2008.

Slides:



Advertisements
Similar presentations
+ Multi-label Classification using Adaptive Neighborhoods Tanwistha Saha, Huzefa Rangwala and Carlotta Domeniconi Department of Computer Science George.
Advertisements

1 Semi-supervised learning for protein classification Brian R. King Chittibabu Guda, Ph.D. Department of Computer Science University at Albany, SUNY Gen*NY*sis.
Semi-supervised Learning Rong Jin. Semi-supervised learning  Label propagation  Transductive learning  Co-training  Active learning.
Large Scale Manifold Transduction Michael Karlen Jason Weston Ayse Erkan Ronan Collobert ICML 2008.
Discriminative Segment Annotation in Weakly Labeled Video Kevin Tang, Rahul Sukthankar Appeared in CVPR 2013 (Oral)
Relational Learning with Gaussian Processes By Wei Chu, Vikas Sindhwani, Zoubin Ghahramani, S.Sathiya Keerthi (Columbia, Chicago, Cambridge, Yahoo!) Presented.
Maria-Florina Balcan Modern Topics in Learning Theory Maria-Florina Balcan 04/19/2006.
Robust Moving Object Detection & Categorization using self- improving classifiers Omar Javed, Saad Ali & Mubarak Shah.
Graph Based Semi- Supervised Learning Fei Wang Department of Statistical Science Cornell University.
Co-Training and Expansion: Towards Bridging Theory and Practice Maria-Florina Balcan, Avrim Blum, Ke Yang Carnegie Mellon University, Computer Science.
Semi-supervised learning and self-training LING 572 Fei Xia 02/14/06.
Semi-Supervised Classification by Low Density Separation Olivier Chapelle, Alexander Zien Student: Ran Chang.
0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Inductive Semi-supervised Learning Gholamreza Haffari Supervised by: Dr. Anoop Sarkar Simon Fraser University, School of Computing Science.
Semi Supervised Learning Qiang Yang –Adapted from… Thanks –Zhi-Hua Zhou – ople/zhouzh/ –LAMDA.
Semi-supervised Learning
Spatial Semi- supervised Image Classification Stuart Ness G07 - Csci 8701 Final Project 1.
MSRC Summer School - 30/06/2009 Cambridge – UK Hybrids of generative and discriminative methods for machine learning.
Combining Labeled and Unlabeled Data for Multiclass Text Categorization Rayid Ghani Accenture Technology Labs.
Semi-Supervised Learning Using Randomized Mincuts Avrim Blum, John Lafferty, Raja Reddy, Mugizi Rwebangira.
Analysis of Semi-supervised Learning with the Yarowsky Algorithm
Co-training LING 572 Fei Xia 02/21/06. Overview Proposed by Blum and Mitchell (1998) Important work: –(Nigam and Ghani, 2000) –(Goldman and Zhou, 2000)
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
Semi-Supervised Learning D. Zhou, O Bousquet, T. Navin Lan, J. Weston, B. Schokopf J. Weston, B. Schokopf Presents: Tal Babaioff.
Bing LiuCS Department, UIC1 Chapter 8: Semi-Supervised Learning Also called “partially supervised learning”
CS Ensembles and Bayes1 Semi-Supervised Learning Can we improve the quality of our learning by combining labeled and unlabeled data Usually a lot.
Maria-Florina Balcan A Theoretical Model for Learning from Labeled and Unlabeled Data Maria-Florina Balcan & Avrim Blum Carnegie Mellon University, Computer.
Using Image Priors in Maximum Margin Classifiers Tali Brayer Margarita Osadchy Daniel Keren.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Semi-supervised Learning Rong Jin. Semi-supervised learning  Label propagation  Transductive learning  Co-training  Active learing.
Diffusion Maps and Spectral Clustering
Incorporating Unlabeled Data in the Learning Process
August 16, 2015EECS, OSU1 Learning with Ambiguously Labeled Training Data Kshitij Judah Ph.D. student Advisor: Prof. Alan Fern Qualifier Oral Presentation.
Semi-Supervised Learning
(ACM KDD 09’) Prem Melville, Wojciech Gryc, Richard D. Lawrence
Interactive Image Segmentation of Non-Contiguous Classes using Particle Competition and Cooperation Fabricio Breve São Paulo State University (UNESP)
Transfer Learning From Multiple Source Domains via Consensus Regularization Ping Luo, Fuzhen Zhuang, Hui Xiong, Yuhong Xiong, Qing He.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
1 Logistic Regression Adapted from: Tom Mitchell’s Machine Learning Book Evan Wei Xiang and Qiang Yang.
Semisupervised Learning A brief introduction. Semisupervised Learning Introduction Types of semisupervised learning Paper for review References.
Ran El-Yaniv and Dmitry Pechyony Technion – Israel Institute of Technology, Haifa, Israel Transductive Rademacher Complexity and its Applications.
1 Learning with Local and Global Consistency Presented by Qiuhua Liu Duke University Machine Learning Group March 23, 2007 By Dengyong Zhou, Olivier Bousquet,
1 COMP3503 Semi-Supervised Learning COMP3503 Semi-Supervised Learning Daniel L. Silver.
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
GRASP Learning a Kernel Matrix for Nonlinear Dimensionality Reduction Kilian Q. Weinberger, Fei Sha and Lawrence K. Saul ICML’04 Department of Computer.
Paired Sampling in Density-Sensitive Active Learning Pinar Donmez joint work with Jaime G. Carbonell Language Technologies Institute School of Computer.
Combining labeled and unlabeled data for text categorization with a large number of categories Rayid Ghani KDD Lab Project.
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
An Overview on Semi-Supervised Learning Methods Matthias Seeger MPI for Biological Cybernetics Tuebingen, Germany.
Bing LiuCS Department, UIC1 Chapter 8: Semi-supervised learning.
COP5992 – DATA MINING TERM PROJECT RANDOM SUBSPACE METHOD + CO-TRAINING by SELIM KALAYCI.
Multiple Instance Learning for Sparse Positive Bags Razvan C. Bunescu Machine Learning Group Department of Computer Sciences University of Texas at Austin.
Lecture 2: Statistical learning primer for biologists
Analysis of Bootstrapping Algorithms Seminar of Machine Learning for Text Mining UPC, 18/11/2004 Mihai Surdeanu.
GeneMANIA: a real-time multiple association network integration algorithm for predicting gene function Sara Mostafavi, Debajyoti Ray, David Warde-Farley,
1 Multi-Task Semi-Supervised Underwater Mine Detection Lawrence Carin, Qiuhua Liu and Xuejun Liao Duke University Jason Stack Office of Naval Research.
CS6772 Advanced Machine Learning Fall 2006 Extending Maximum Entropy Discrimination on Mixtures of Gaussians With Transduction Final Project by Barry.
NTU & MSRA Ming-Feng Tsai
Classification using Co-Training
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
1 Boosting Algorithms for Supervised Learning and Semi-supervised Learning Hongbo Deng RM :30 am – 12:00 May 7, 2007 PhD. Term Presentation, CSE.
Graph-based WSD の続き DMLA /7/10 小町守.
Introduction to Machine Learning Nir Ailon Lecture 11: Probabilistic Models.
Semi-Supervised Learning Using Label Mean
Semi-Supervised Clustering
Graph Based Multi-Modality Learning
Generally Discriminant Analysis
Multivariate Methods Berlin Chen
Basics of ML Rohan Suri.
Multivariate Methods Berlin Chen, 2005 References:
Presentation transcript:

Graph-Based Semi-Supervised Learning with a Generative Model Speaker: Jingrui He Advisor: Jaime Carbonell Machine Learning Department

04/03/20082 Semi-Supervised Learning - + Very Few Abundant

04/03/20083 Outline ► Background ► Existing Methods ► Proposed Method  Ideal Case  General Case ► Experimental Results ► Conclusion

04/03/20084 Overview Semi- Supervised Learning Feature based Graph based Gradually generate class labels Collectively generate class labels Mincut [Blum, ICML01] Gaussian Random Fields [Zhu, ICML03] Local and Global Consistency [Zhou, NIPS04] Generative Model [He, IJCAI07] Self-Training, [Yarowsky, ACL95] Co-Training, [Blum, COLT98] TSVMs [Joachims, ICML99] EM-based, [Nigam, ML00]

04/03/20085 Self-Training [Yarowsky, ACL95] - +

04/03/20086 Co-Training [Blum, COLT98] Sufficient to train a good classifier Conditionally independent given the class

04/03/20087 Transductive SVMs [Joachims, ICML99] - + Inductive SVMs Transductive SVMs Classification Boundary: Away from the Dense Regions!

04/03/20088 EM-based Method [Nigam, ML00] Text Corpus Computer Science Medicine Politics

04/03/ Graph-Based Semi-Supervised Learning

04/03/ Graph-Based Methods ► G={V,E} ► Estimating a function f on the graph  f should be close to the given labels on the labeled nodes  f should be smooth on the whole graph ► Regularization

04/03/ Graph-Based Methods cont. ► Mincut [Blum, ICML01]  ► Gaussian Random Fields [Zhu, ICML03]  ► Local and Global Consistency [Zhou, NIPS04]  ► Discriminative in Nature!

04/03/ Outline ► Background ► Existing Methods ► Proposed Method  Ideal Case  General Case ► Experimental Results ► Conclusion

04/03/ Motivation ► Existing Graph-Based Methods:  : NO justification  Discriminative: inaccurate proportion in the labeled set greatly AFFECTS the performance ► Proposed Method:  : WELL justified  Generative: estimated class priors COMPENSATES for the inaccurate proportion in the labeled set

04/03/ Notation ► n training examples: ► labeled examples, ► unlabeled examples ► Affinity matrix: ► similarity between and ► Diagonal matrix D : ► ► : set to 1 for labeled examples

04/03/ Ideal Case ► Two classes far apart

04/03/ Derivation Sketch Relate to Relate eigenvector to Relate to

04/03/ Class Conditional Probability ► Theorem 1  As,  Similar to kernel density estimation ► Unlabeled data  ? ?

04/03/ Class Conditional Probability cont. ► Eigenvectors of S  ; ► Element-wise: ► ;

04/03/ Class Conditional Probability cont. ► To get and, iterate: , ► Upon convergence , ► After normalization ,

04/03/ Example of the Ideal Case

04/03/ General Case ► Two classes not far apart ► S not block diagonal Upon Convergence

04/03/ Class Conditional Probability ► Iteration process  The labeled examples gradually spread their information to nearby points ► Solution  Stop the iteration when certain criterion is satisfied

04/03/ Stopping Criterion ► Average probability of the negative labeled examples in the positive class

04/03/ Stopping Criterion cont. Pre- maturity Excessive Propagation

04/03/ Stopping Criterion cont. ► Average probability of the positive labeled examples in the negative class

04/03/ Example of the General Case

04/03/ Estimating Class Priors ► Theorem 2: in the general case, as  ► To get estimates of  

04/03/ Prediction ► To classify a new example  Calculate the class conditional probabilities  According to Bayes rule

04/03/ Outline ► Background ► Existing Methods ► Proposed Method  Ideal Case  General Case ► Experimental Results ► Conclusion

04/03/ Cedar Buffalo Binary Digits Data Set [Hull, PAMI94] ► Balanced classification 1 vs 2odd vs even Our method Gaussian Random Fields Local and Global Consistency Our method Gaussian Random Fields Local and Global Consistency

04/03/ Cedar Buffalo Binary Digits Data Set [Hull, PAMI94] ► Unbalanced classification Our method Gaussian Random Fields Local and Global Consistency Our method Gaussian Random Fields Local and Global Consistency 1 vs 2odd vs even

04/03/ Genre Data Set [Liu, ECML03] ► Classification between random partitions balancedunbalanced Our method Gaussian Random Fields Local and Global Consistency Our method Gaussian Random Fields Local and Global Consistency

04/03/ Genre Data Set [Liu, ECML03] ► Unbalanced classification newspapers vs otherbiographies vs other Our method Gaussian Random Fields Local and Global Consistency Our method Gaussian Random Fields Local and Global Consistency

04/03/ Conclusion ► A new graph-based semi-supervised learning method  Generative in nature  Ideal case: theoretical guarantee  General case: reasonable estimates  Prediction: easy and intuitive

Questions?