Open-Category Classification by Adversarial Sample Generation

Slides:



Advertisements
Similar presentations
January 23 rd, Document classification task We are interested to solve a task of Text Classification, i.e. to automatically assign a given document.
Advertisements

Mustafa Cayci INFS 795 An Evaluation on Feature Selection for Text Clustering.
ICONIP 2005 Improve Naïve Bayesian Classifier by Discriminative Training Kaizhu Huang, Zhangbing Zhou, Irwin King, Michael R. Lyu Oct
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
IMAN SAUDY UMUT OGUR NORBERT KISS GEORGE TEPES-NICA BARLEY SEEDS CLASSIFICATION.
Learning on Probabilistic Labels Peng Peng, Raymond Chi-wing Wong, Philip S. Yu CSE, HKUST 1.
Software Quality Ranking: Bringing Order to Software Modules in Testing Fei Xing Michael R. Lyu Ping Guo.
Prénom Nom Document Analysis: Linear Discrimination Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Logistic Regression Rong Jin. Logistic Regression Model  In Gaussian generative model:  Generalize the ratio to a linear model Parameters: w and c.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Classification with reject option in gene expression data Blaise Hanczar and Edward R Dougherty BIOINFORMATICS Vol. 24 no , pages
Parameterizing Random Test Data According to Equivalence Classes Chris Murphy, Gail Kaiser, Marta Arias Columbia University.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Active Learning for Class Imbalance Problem
Cost-Sensitive Bayesian Network algorithm Introduction: Machine learning algorithms are becoming an increasingly important area for research and application.
1 Logistic Regression Adapted from: Tom Mitchell’s Machine Learning Book Evan Wei Xiang and Qiang Yang.
DATA MINING LECTURE 10 Classification k-nearest neighbor classifier Naïve Bayes Logistic Regression Support Vector Machines.
An Introduction to Support Vector Machine (SVM) Presenter : Ahey Date : 2007/07/20 The slides are based on lecture notes of Prof. 林智仁 and Daniel Yeung.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
A Novel Local Patch Framework for Fixing Supervised Learning Models Yilei Wang 1, Bingzheng Wei 2, Jun Yan 2, Yang Hu 2, Zhi-Hong Deng 1, Zheng Chen 2.
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
An Introduction to Support Vector Machine (SVM)
Class Imbalance Classification Implementation Group 4 WEI Lili, ZENG Gaoxiong,
Jen-Tzung Chien, Meng-Sung Wu Minimum Rank Error Language Modeling.
Iterative similarity based adaptation technique for Cross Domain text classification Under: Prof. Amitabha Mukherjee By: Narendra Roy Roll no: Group:
Improving Support Vector Machine through Parameter Optimized Rujiang Bai, Junhua Liao Shandong University of Technology Library Zibo , China { brj,
Locally Linear Support Vector Machines Ľubor Ladický Philip H.S. Torr.
Spoken Language Group Chinese Information Processing Lab. Institute of Information Science Academia Sinica, Taipei, Taiwan
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
Linear Discriminant Functions Chapter 5 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis.
Neural networks and support vector machines
Support vector machines
Bridging Domains Using World Wide Knowledge for Transfer Learning
Robust and Fast Collaborative Tracking with Two Stage Sparse Optimization Authors: Baiyang Liu, Lin Yang, Junzhou Huang, Peter Meer, Leiguang Gong and.
Semi-Supervised Clustering
Evaluating Classifiers
Deep Feedforward Networks
Introduction to Machine Learning
Decision Trees.
Presented by Jingting Zeng 11/26/2007
Data Driven Attributes for Action Detection
KDD 2004: Adversarial Classification
Data Mining Lecture 11.
Importance Weighted Active Learning
LINEAR AND NON-LINEAR CLASSIFICATION USING SVM and KERNELS
Machine Learning Week 1.
Privacy-Preserving Classification
A weight-incorporated similarity-based clustering ensemble method based on swarm intelligence Yue Ming NJIT#:
Learning with information of features
Outline Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no.
Statistical Learning Dong Liu Dept. EEIS, USTC.
Ying shen Sse, tongji university Sep. 2016
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Generative Models and Naïve Bayes
Discriminative Frequent Pattern Analysis for Effective Classification
COSC 4335: Other Classification Techniques
Supervised vs. unsupervised Learning
Shih-Wei Lin, Kuo-Ching Ying, Shih-Chieh Chen, Zne-Jung Lee
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Xin Qi, Matthew Keally, Gang Zhou, Yantao Li, Zhen Ren
Support vector machines
Zhedong Zheng, Liang Zheng and Yi Yang
Machine Learning: Lecture 6
Machine Learning: UNIT-3 CHAPTER-1
Linear Discrimination
Supervised machine learning: creating a model
Discriminative Training
Learning to Navigate for Fine-grained Classification
Presentation transcript:

Open-Category Classification by Adversarial Sample Generation Yang Yu, Wei-Yang Qu, Nan Li, and Zimin Guo 汇报人:刘驭壬

Outline Introduction Background Proposed Method Experiments Conclusion Open-Category Classification Background Related Problems Adversarial Learning Derivative-free Optimization Proposed Method ASG Framework Experiments Conclusion

Open-Category Classification An example Open-Category Classification there are novel classes that none of their instances are observed during the training phase, but in the test phase their instances could be encountered

Open-Category Classification

Related Problems Incremental Learning Learning with Rejection Option New classes are assumed to appear incremental in the training set Learning with Rejection Option Classifier reject to recognize an instance if its confidence is low Outlier Detection Apply to OCC problem by treating unseen class instances as outliers Zero-Shot Learning commonly assume that a high-level attribute set is available for all classes including the unseen one

Adversarial Learning Adversarial Learning model Discriminative Model: Try to discriminate the origin data and generated data Generative Model: Learns to generate instances that can fool the discriminative model as a non-generated instance Generative Adversarial Nets(GAN) result in a model that is consistent with the original data distribution

Derivative-free Optimization A derivative-free optimization method consider an optimization formalized as: Instead of calculating gradients of , samples solutions and learns from their feedbacks for finding better solutions Some Derivative-free Methods genetic algorithms, most are heuristic methods Bayesian optimization methods Optimistic optimization methods Model-based optimization(i.e. RACOS)

ASG Framework Minimize: Consideration: No instances of class novel Instances considered to belong to the seen class Instances considered to belong to the unseen class Generate instances of class novel Generate positive and negative instances of class seen

Generate negative instances of class seen Goal: optimize a point in the instance space to be close to the seen class instances, but cannot be recognized as an seen class instance by the discriminator Optimize:

Generate negative instances of class seen Optimize: Regular term Limit the distance of points belongs to and The distance of points in should not be too small Minimize:

Generate negative instances of class seen Algorithm T is the numbers to be sampled in practice Eq.(8):

Generate positive instances of class seen Optimize: Regular term The distance of points in should not be too small Minimize: Algorithm: just replace Eq.(8) with Eq.(11) in Algorithm.1

Experiments Discriminant model: RBF-SVM Compared methods: OC-SVM, MOC-SVM, 1-vs-Set Machine, OVR-SVM

Experiment Results Results on 3 moon: Results on MNIST:

Experiment Results Results on MNIST: Results on the 20 News Group: Results on the 5 small datasets:

Conclusion Open-category classification problem is a practical problem when a system needs to predict the data in open environments propose to address the problem by adversarial data generation Experiments on several datasets show the effectiveness of the proposed method

Thanks!