Bing LiuCS Department, UIC1 Chapter 8: Semi-supervised learning.

Slides:



Advertisements
Similar presentations
PEBL: Web Page Classification without Negative Examples Hwanjo Yu, Jiawei Han, Kevin Chen- Chuan Chang IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING,
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Support Vector Machine
Sequential Minimal Optimization Advanced Machine Learning Course 2012 Fall Semester Tsinghua University.
ECG Signal processing (2)
An Introduction of Support Vector Machine
Chapter 5: Partially-Supervised Learning
Chapter 7: Text mining UIC - CS 594 Bing Liu 1 1.
Prénom Nom Document Analysis: Linear Discrimination Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Optimizing F-Measure with Support Vector Machines David R. Musicant Vipin Kumar Aysel Ozgur FLAIRS 2003 Tuesday, May 13, 2003 Carleton College.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Spatial Semi- supervised Image Classification Stuart Ness G07 - Csci 8701 Final Project 1.
Sample Selection Bias Lei Tang Feb. 20th, Classical ML vs. Reality  Training data and Test data share the same distribution (In classical Machine.
Active Learning with Support Vector Machines
Support Vector Machines Kernel Machines
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
Dept. of Computer Science & Engineering, CUHK Pseudo Relevance Feedback with Biased Support Vector Machine in Multimedia Retrieval Steven C.H. Hoi 14-Oct,
Bing LiuCS Department, UIC1 Learning from Positive and Unlabeled Examples Bing Liu Department of Computer Science University of Illinois at Chicago Joint.
Bing LiuCS Department, UIC1 Chapter 8: Semi-Supervised Learning Also called “partially supervised learning”
Linear Discriminant Functions Chapter 5 (Duda et al.)
Greg GrudicIntro AI1 Support Vector Machine (SVM) Classification Greg Grudic.
Final review LING572 Fei Xia Week 10: 03/11/
Learning from Imbalanced, Only Positive and Unlabeled Data Yetian Chen
Active Learning for Class Imbalance Problem
Learning with Positive and Unlabeled Examples using Weighted Logistic Regression Wee Sun Lee National University of Singapore Bing Liu University of Illinois,
SVM by Sequential Minimal Optimization (SMO)
CS 8751 ML & KDDSupport Vector Machines1 Support Vector Machines (SVMs) Learning mechanism based on linear programming Chooses a separating plane based.
Machine Learning CSE 681 CH2 - Supervised Learning.
Classification and Ranking Approaches to Discriminative Language Modeling for ASR Erinç Dikici, Murat Semerci, Murat Saraçlar, Ethem Alpaydın 報告者:郝柏翰 2013/01/28.
Universit at Dortmund, LS VIII
Partially Supervised Classification of Text Documents by Bing Liu, Philip Yu, and Xiaoli Li Presented by: Rick Knowles 7 April 2005.
ICML2004, Banff, Alberta, Canada Learning Larger Margin Machine Locally and Globally Kaizhu Huang Haiqin Yang, Irwin King, Michael.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
SVM Support Vector Machines Presented by: Anas Assiri Supervisor Prof. Dr. Mohamed Batouche.
Machine Learning in Ad-hoc IR. Machine Learning for ad hoc IR We’ve looked at methods for ranking documents in IR using factors like –Cosine similarity,
Greedy is not Enough: An Efficient Batch Mode Active Learning Algorithm Chen, Yi-wen( 陳憶文 ) Graduate Institute of Computer Science & Information Engineering.
Web Mining ( 網路探勘 ) WM05 TLMXM1A Wed 8,9 (15:10-17:00) U705 Partially Supervised Learning ( 部分監督式學習 ) Min-Yuh Day 戴敏育 Assistant Professor 專任助理教授.
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
Learning from Positive and Unlabeled Examples Investigator: Bing Liu, Computer Science Prime Grant Support: National Science Foundation Problem Statement.
CSSE463: Image Recognition Day 14 Lab due Weds, 3:25. Lab due Weds, 3:25. My solutions assume that you don't threshold the shapes.ppt image. My solutions.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Supervised Learning. CS583, Bing Liu, UIC 2 An example application An emergency room in a hospital measures 17 variables (e.g., blood pressure, age, etc)
Support Vector Machines
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Comparative Experiments on Sentiment Classification for Online Product Reviews Hang Cui, Vibhu Mittal, and Mayur Datar AAAI 2006.
Classification Course web page: vision.cis.udel.edu/~cv May 14, 2003  Lecture 34.
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
Univariate Gaussian Case (Cont.)
Machine Learning in Practice Lecture 10 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
6.S093 Visual Recognition through Machine Learning Competition Image by kirkh.deviantart.com Joseph Lim and Aditya Khosla Acknowledgment: Many slides from.
Support Vector Machines Reading: Ben-Hur and Weston, “A User’s Guide to Support Vector Machines” (linked from class web page)
 Effective Multi-Label Active Learning for Text Classification Bishan yang, Juan-Tao Sun, Tengjiao Wang, Zheng Chen KDD’ 09 Supervisor: Koh Jia-Ling Presenter:
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Bayes Rule Mutual Information Conditional.
High resolution product by SVM. L’Aquila experience and prospects for the validation site R. Anniballe DIET- Sapienza University of Rome.
Unsupervised Learning Part 2. Topics How to determine the K in K-means? Hierarchical clustering Soft clustering with Gaussian mixture models Expectation-Maximization.
Partially Supervised Classification of Text Documents
Semi-Supervised Clustering
LINEAR CLASSIFIERS The Problem: Consider a two class task with ω1, ω2.
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Constrained Clustering -Semi Supervised Clustering-
Chapter 8: Semi-Supervised Learning
LINEAR AND NON-LINEAR CLASSIFICATION USING SVM and KERNELS
PEBL: Web Page Classification without Negative Examples
Unsupervised Learning II: Soft Clustering with Gaussian Mixture Models
LECTURE 09: BAYESIAN LEARNING
Concave Minimization for Support Vector Machine Classifiers
MAS 622J Course Project Classification of Affective States - GP Semi-Supervised Learning, SVM and kNN Hyungil Ahn
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Presentation transcript:

Bing LiuCS Department, UIC1 Chapter 8: Semi-supervised learning

Bing LiuCS Department, UIC2 Introduction There are mainly two types of semi- supervised learning, or partially supervised learning. Learning with positive and unlabeled training set (no labeled negative data). Learning with a small labeled training set of every classes and a large unlabeled set. We first consider Learning with positive and unlabeled training set.

Bing LiuCS Department, UIC3 Classic Supervised Learning Given a set of labeled training examples of n classes, the system uses this set to build a classifier. The classifier is then used to classify new data into the n classes. Although this traditional model is very useful, in practice one also encounters another (related) problem.

Bing LiuCS Department, UIC4 Learning from Positive & Unlabeled data (PU-learning) Positive examples: One has a set of examples of a class P, and Unlabeled set: also has a set U of unlabeled (or mixed) examples with instances from P and also not from P (negative examples). Build a classifier: Build a classifier to classify the examples in U and/or future (test) data. Key feature of the problem: no labeled negative training data. We call this problem, PU-learning.

Bing LiuCS Department, UIC5 Applications of the problem With the growing volume of online texts available through the Web and digital libraries, one often wants to find those documents that are related to one's work or one's interest. For example, given a ICML proceedings, find all machine learning papers from AAAI, IJCAI, KDD No labeling of negative examples from each of these collections. Similarly, given one's bookmarks (positive documents), identify those documents that are of interest to him/her from Web sources.

Bing LiuCS Department, UIC6 Are Unlabeled Examples Helpful? Function known to be either x 1 0 Which one is it? x 1 < 0 x 2 > u u u u u u u u u u u “Not learnable” with only positive examples. However, addition of unlabeled examples makes it learnable.

Bing LiuCS Department, UIC7 Theoretical foundations (X, Y): X - input vector, Y  {1, -1} - class label. f : classification function We rewrite the probability of error Pr[f(X)  Y] = Pr[f(X) = 1 and Y = -1] + (1) Pr[f(X) = -1 and Y = 1] We have Pr[f(X) = 1 and Y = -1] = Pr[f(X) = 1] – Pr[f(X) = 1 and Y = 1] = Pr[f(X) = 1] – (Pr[Y = 1] – Pr[f(X) = -1 and Y = 1]). Plug this into (1), we obtain Pr[f(X)  Y] = Pr[f(X) = 1] – Pr[Y = 1] (2) + 2Pr[f(X) = -1|Y = 1]Pr[Y = 1]

Bing LiuCS Department, UIC8 Theoretical foundations (cont) Pr[f(X)  Y] = Pr[f(X) = 1] – Pr[Y = 1] (2) + 2Pr[f(X) = -1|Y = 1] Pr[Y = 1] Note that Pr[Y = 1] is constant. If we can hold Pr[f(X) = -1|Y = 1] small, then learning is approximately the same as minimizing Pr[f(X) = 1]. Holding Pr[f(X) = -1|Y = 1] small while minimizing Pr[f(X) = 1] is approximately the same as minimizing Pr u [f(X) = 1] while holding Pr P [f(X) = 1] ≥ r (where r is recall) if the set of positive examples P and the set of unlabeled examples U are large enough.

Bing LiuCS Department, UIC9 Put it simply A constrained optimization problem. A reasonably good generalization (learning) result can be achieved If the algorithm tries to minimize the number of unlabeled examples labeled as positive subject to the constraint that the fraction of errors on the positive examples is no more than 1-r.

Bing LiuCS Department, UIC10 Existing 2-step strategy Step 1: Identifying a set of reliable negative documents from the unlabeled set. Step 2: Building a sequence of classifiers by iteratively applying a classification algorithm and then selecting a good classifier.

Bing LiuCS Department, UIC11 Step 1: The Spy technique Sample a certain % of positive examples and put them into unlabeled set to act as “spies”. Run a classification algorithm assuming all unlabeled examples are negative, we will know the behavior of those actual positive examples in the unlabeled set through the “spies”. We can then extract reliable negative examples from the unlabeled set more accurately.

Bing LiuCS Department, UIC12 Step 1: Other methods 1-DNF method: Find the set of words W that occur in the positive documents more frequently than in the unlabeled set. Extract those documents from unlabeled set that do not contain any word in W. These documents form the reliable negative documents. Rocchio method from information retrieval Naïve Bayesian method.

Bing LiuCS Department, UIC13 Step 2: Running EM or SVM iteratively (1) Running a classification algorithm iteratively Run EM using P, RN and Q until it converges, or Run SVM iteratively using P, RN and Q until this no document from Q can be classified as negative. RN and Q are updated in each iteration, or … (2) Classifier selection.

Bing LiuCS Department, UIC14 Do they follow the theory? Yes, heuristic methods because Step 1 tries to find some initial reliable negative examples from the unlabeled set. Step 2 tried to identify more and more negative examples iteratively. The two steps together form an iterative strategy of increasing the number of unlabeled examples that are classified as negative while maintaining the positive examples correctly classified.

Bing LiuCS Department, UIC15 Can SVM be applied directly? Can we use SVM to directly deal with the problem of learning with positive and unlabeled examples, without using two steps? Yes, with a little re-formulation. The theory says that if the sample size is large enough, minimizing the number of unlabeled examples classified as positive while constraining the positive examples to be correctly classified will give a good classifier.

Bing LiuCS Department, UIC16 Support Vector Machines Support vector machines (SVM) are linear functions of the form f(x) = w T x + b, where w is the weight vector and x is the input vector. Let the set of training examples be {(x 1, y 1 ), (x 2, y 2 ), …, (x n, y n )}, where x i is an input vector and y i is its class label, y i  {1, -1}. To find the linear function: Minimize: Subject to:

Bing LiuCS Department, UIC17 Soft margin SVM To deal with cases where there may be no separating hyperplane due to noisy labels of both positive and negative training examples, the soft margin SVM is proposed: Minimize: Subject to: where C  0 is a parameter that controls the amount of training errors allowed.

Bing LiuCS Department, UIC18 Biased SVM (noiseless case) Assume that the first k-1 examples are positive examples (labeled 1), while the rest are unlabeled examples, which we label negative (-1). Minimize: Subject to:  i  0, i = k, k+1…, n

Bing LiuCS Department, UIC19 Biased SVM (noisy case) If we also allow positive set to have some noisy negative examples, then we have: Minimize: Subject to:  i  0, i = 1, 2, …, n. This turns out to be the same as the asymmetric cost SVM for dealing with unbalanced data. Of course, we have a different motivation.

Bing LiuCS Department, UIC20 Estimating performance We need to estimate the performance in order to select the parameters. Since learning from positive and negative examples often arise in retrieval situations, we use F score as the classification performance measure F = 2pr / (p+r) (p: precision, r: recall). To get a high F score, both precision and recall have to be high. However, without labeled negative examples, we do not know how to estimate the F score.

Bing LiuCS Department, UIC21 A performance criterion Performance criteria pr/Pr[Y=1]: It can be estimated directly from the validation set as r 2 /Pr[f(X) = 1] Recall r = Pr[f(X)=1| Y=1] Precision p = Pr[Y=1| f(X)=1] To see this Pr[f(X)=1|Y=1] Pr[Y=1] = Pr[Y=1|f(X)=1] Pr[f(X)=1]  // both side times r Behavior similar to the F-score (= 2pr / (p+r))

Bing LiuCS Department, UIC22 A performance criterion (cont …) r 2 /Pr[f(X) = 1] r can be estimated from positive examples in the validation set. Pr[f(X) = 1] can be obtained using the full validation set. This criterion actually reflects our theory very well.

Bing LiuCS Department, UIC23 Summary Gave an overview of the theory on learning with positive and unlabeled examples. Described the existing two-step strategy for learning. Presented an more principled approach to solve the problem based on a biased SVM formulation. Presented a performance measure pr/P(Y=1) that can be estimated from data.