On the Hardness of Evading Combinations of Linear Classifiers Daniel Lowd University of Oregon Joint work with David Stevens.

Slides:



Advertisements
Similar presentations
Statistical Machine Learning- The Basic Approach and Current Research Challenges Shai Ben-David CS497 February, 2007.
Advertisements

Lecture 24 MAS 714 Hartmut Klauck
Dana Nau: Lecture slides for Automated Planning Licensed under the Creative Commons Attribution-NonCommercial-ShareAlike License:
Integrated Instance- and Class- based Generative Modeling for Text Classification Antti PuurulaUniversity of Waikato Sung-Hyon MyaengKAIST 5/12/2013 Australasian.
Imbalanced data David Kauchak CS 451 – Fall 2013.
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
SVM—Support Vector Machines
Machine Learning Theory Machine Learning Theory Maria Florina Balcan 04/29/10 Plan for today: - problem of “combining expert advice” - course retrospective.
Eran Omri, Bar-Ilan University Joint work with Amos Beimel and Ilan Orlov, BGU Ilan Orlov…!??!!
Foundations of Adversarial Learning Daniel Lowd, University of Washington Christopher Meek, Microsoft Research Pedro Domingos, University of Washington.
Partitioned Logistic Regression for Spam Filtering Ming-wei Chang University of Illinois at Urbana-Champaign Wen-tau Yih and Christopher Meek Microsoft.
1 s-t Graph Cuts for Binary Energy Minimization  Now that we have an energy function, the big question is how do we minimize it? n Exhaustive search is.
A Fairy Tale of Greedy Algorithms Yuli Ye Joint work with Allan Borodin, University of Toronto.
Models and Security Requirements for IDS. Overview The system and attack model Security requirements for IDS –Sensitivity –Detection Analysis methodology.
Yi Wu (CMU) Joint work with Vitaly Feldman (IBM) Venkat Guruswami (CMU) Prasad Ragvenhdra (MSR)
CSE 326: Data Structures NP Completeness Ben Lerner Summer 2007.
Analysis of Algorithms CS 477/677
Job Scheduling Lecture 19: March 19. Job Scheduling: Unrelated Multiple Machines There are n jobs, each job has: a processing time p(i,j) (the time to.
Adversarial Learning: Practice and Theory Daniel Lowd University of Washington July 14th, 2006 Joint work with Chris Meek, Microsoft Research “If you know.
Foundations of Adversarial Learning Daniel Lowd, University of Washington Christopher Meek, Microsoft Research Pedro Domingos, University of Washington.
(work appeared in SODA 10’) Yuk Hei Chan (Tom)
Linear Discriminators Chapter 20 From Data to Knowledge.
On Solving Presburger and Linear Arithmetic with SAT Ofer Strichman Carnegie Mellon University.
Machine Learning Theory Maria-Florina Balcan Lecture 1, Jan. 12 th 2010.
Foundations of Cryptography Lecture 2 Lecturer: Moni Naor.
Online Learning Algorithms
Good Word Attacks on Statistical Spam Filters Daniel Lowd University of Washington (Joint work with Christopher Meek, Microsoft Research)
Machine Learning Theory Maria-Florina (Nina) Balcan Lecture 1, August 23 rd 2011.
Strategy-Proof Classification Reshef Meir School of Computer Science and Engineering, Hebrew University A joint work with Ariel. D. Procaccia and Jeffrey.
Masquerade Detection Mark Stamp 1Masquerade Detection.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
Security Evaluation of Pattern Classifiers under Attack.
Trust-Aware Optimal Crowdsourcing With Budget Constraint Xiangyang Liu 1, He He 2, and John S. Baras 1 1 Institute for Systems Research and Department.
Man vs. Machine: Adversarial Detection of Malicious Crowdsourcing Workers Gang Wang, Tianyi Wang, Haitao Zheng, Ben Y. Zhao, UC Santa Barbara, Usenix Security.
Hardness of Learning Halfspaces with Noise Prasad Raghavendra Advisor Venkatesan Guruswami.
Constraint Satisfaction Problems (CSPs) CPSC 322 – CSP 1 Poole & Mackworth textbook: Sections § Lecturer: Alan Mackworth September 28, 2012.
Disclosure risk when responding to queries with deterministic guarantees Krish Muralidhar University of Kentucky Rathindra Sarathy Oklahoma State University.
Improving Cloaking Detection Using Search Query Popularity and Monetizability Kumar Chellapilla and David M Chickering Live Labs, Microsoft.
Support Vector Machines Reading: Ben-Hur and Weston, “A User’s Guide to Support Vector Machines” (linked from class web page)
Decision Diagrams for Sequencing and Scheduling Andre Augusto Cire Joint work with David Bergman, Willem-Jan van Hoeve, and John Hooker Tepper School of.
Deformable Part Models (DPM) Felzenswalb, Girshick, McAllester & Ramanan (2010) Slides drawn from a tutorial By R. Girshick AP 12% 27% 36% 45% 49% 2005.
1 The Theory of NP-Completeness 2 Cook ’ s Theorem (1971) Prof. Cook Toronto U. Receiving Turing Award (1982) Discussing difficult problems: worst case.
Bahman Bahmani Stanford University
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
CSSE463: Image Recognition Day 14 Lab due Weds, 3:25. Lab due Weds, 3:25. My solutions assume that you don't threshold the shapes.ppt image. My solutions.
Linear Programming Maximize Subject to Worst case polynomial time algorithms for linear programming 1.The ellipsoid algorithm (Khachian, 1979) 2.Interior.
CZ5225: Modeling and Simulation in Biology Lecture 7, Microarray Class Classification by Machine learning Methods Prof. Chen Yu Zong Tel:
Support Vector Machines. Notation Assume a binary classification problem. –Instances are represented by vector x   n. –Training examples: x = (x 1,
Machine Learning Concept Learning General-to Specific Ordering
Donghyun (David) Kim Department of Mathematics and Computer Science North Carolina Central University 1 Chapter 7 Time Complexity Some slides are in courtesy.
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
6.S093 Visual Recognition through Machine Learning Competition Image by kirkh.deviantart.com Joseph Lim and Aditya Khosla Acknowledgment: Many slides from.
Support Vector Machines Reading: Ben-Hur and Weston, “A User’s Guide to Support Vector Machines” (linked from class web page)
Common Intersection of Half-Planes in R 2 2 PROBLEM (Common Intersection of half- planes in R 2 ) Given n half-planes H 1, H 2,..., H n in R 2 compute.
Computer Systems Laboratory Stanford University Clark W. Barrett David L. Dill Aaron Stump A Framework for Cooperating Decision Procedures.
On-Line Algorithms in Machine Learning By: WALEED ABDULWAHAB YAHYA AL-GOBI MUHAMMAD BURHAN HAFEZ KIM HYEONGCHEOL HE RUIDAN SHANG XINDI.
Page 1 CS 546 Machine Learning in NLP Review 1: Supervised Learning, Binary Classifiers Dan Roth Department of Computer Science University of Illinois.
Theory of Computational Complexity Yusuke FURUKAWA Iwama Ito lab M1.
Unconstrained Submodular Maximization Moran Feldman The Open University of Israel Based On Maximizing Non-monotone Submodular Functions. Uriel Feige, Vahab.
1 Bilinear Classifiers for Visual Recognition Computational Vision Lab. University of California Irvine To be presented in NIPS 2009 Hamed Pirsiavash Deva.
1 CS 391L: Machine Learning: Computational Learning Theory Raymond J. Mooney University of Texas at Austin.
Dan Roth Department of Computer and Information Science
Introduction to Machine Learning
Dude, where’s that IP? Circumventing measurement-based geolocation
Phillipa Gill University of Toronto
Adversarial Evasion-Resilient Hardware Malware Detectors
RHMD: Evasion-Resilient Hardware Malware Detectors
Chapter 11 Limitations of Algorithm Power
Binghui Wang, Le Zhang, Neil Zhenqiang Gong
Clustering.
Presentation transcript:

On the Hardness of Evading Combinations of Linear Classifiers Daniel Lowd University of Oregon Joint work with David Stevens

Machine learning is used more and more in adversarial domains… Intrusion detection Malware detection Phishing detection Detecting malicious advertisements Detecting fake reviews Credit card fraud detection Online auction fraud detection spam filtering Blog spam filtering OSN spam filtering Political censorship …and more every year!

Evasion Attack 1.System designers deploy a classifier. 2.An attacker learns about the model through interaction (and possibly other information sources). 3.An attacker uses this knowledge to evade detection by changing its behavior as little as possible. Example: A spammer sends test s to learn how to modify a spam so that it gets past a spam filter. Question: How easily can the attacker learn enough to mount an effective attack?

4 Adversarial Classifier Reverse Engineering (ACRE) Task: Find the negative instance “closest” to x a (We will also refer to this distance as a “cost” to be minimized.) Problem: the adversary doesn ’ t know the classifier! [Lowd&Meek,’05] X1X1 X2X2 + - xaxa

5 Adversarial Classifier Reverse Engineering (ACRE) Task: Find the negative instance “closest” to x a Given: X1X1 X2X2 ?? ? ? ? ? ? ? - + – One positive and one negative instance, x + and x  – A polynomial number of membership queries Within a factor of k [Lowd&Meek,’05] xaxa

Example: Linear Classifiers With continuous features and L 1 distance, find optimal point by doing line search in each dimension: However, with binary features, we can’t do line searches. X1X1 X2X2 xaxa * -- Somewhat more efficient methods exist for the continuous case. [Nelson&al.,2012].

7 Attacking Linear Classifiers with Boolean features Can efficiently find an evasion with at most twice the optimal cost, assuming unit cost for each “change”. xaxa x-x- wiwi wjwj wkwk wlwl wmwm c(x)c(x) METHOD: Iteratively reduce cost in two ways: 1.Remove any unnecessary change: O(n) 2.Replace any two changes with one: O(n 3 ) xaxa y wiwi wjwj wkwk wlwl c(x)c(x) wmwm x-x- xaxa y’y’ wiwi wjwj wkwk wlwl c(x)c(x) wpwp Also known: Any convex-inducing classifier with continuous features is ACRE-learnable. [Nelson&al.,2012] [Lowd&Meek’05]

This work: We consider when the positive or negative class is an intersection of half-spaces, or polytope, representable by combinations of linear classifiers: What about non-linear classifiers? Positive class is conjunction of linear classifiers. Example: One classifier to identify each legitimate user. Positive class is disjunction of linear classifiers. Example: One classifier for each type of attack. We show that the attack problem is hard in general, but easy when the half-spaces are defined over disjoint features.

Hardness Results With continuous features and L 1 costs, near-optimal evasion of a polytope requires polynomially many queries. [Nelson et al., 2012] With discrete features, we show that exponentially many queries are required in the worst case. Proofs work for any fixed approximation ratio k. Key Idea: Construct a set of component classifiers so there is no clear path from “distant” to “close” negative instances.

Hardness of Evading Disjunctions n/2k classifiers n/2+1 Two ways to evade: – Include all light-green features (cost: n/2+1) – Include all dark-green features (cost: n/2k) Challenge: – If you don’t guess all dark-green features, some classifier remains positive. – If you include extra red features, all classifiers become positive. Guessing low-cost instance requires exponentially many queries! (Instance is negative only if all component classifiers mark it as negative.)

Hardness of Evading Conjunctions To evade c 2 : Include > ½ the light-green features (cost: n/4+1) To evade c 1 : Include all dark-green features (cost: n/4k), or all light-green features (cost: n/2), or a combo. Two cases: – When > ½ the light-green features are included, c 2 is negative so dark-greens have no effect on the class label. – When ½ the dark-green features to evade c 1. Adversary must guess n/8k features! n/2 n/4k (Instance is negative only if any component classifier marks it as negative.) c1c1 c2c2

Restriction: Disjoint Features In practice, classifiers do not always represent the worst case. In some applications, each classifier in the set works on a different set of features: – Image or fingerprint biometrics classifiers – Separate image spam and HTML spam classifiers This simple restriction makes attacks easy!

Evading Disjoint Disjunctions Theorem: Linear attack from [Lowd&Meek,2005] is at most twice optimal on disjoint disjunctions. Proof Sketch: When features are disjoint, the optimal evasion is to evade each component classifier optimally. When the algorithm terminates, there is no way to reduce the cost with individual or pairs of changes, so each separate evasion is at most twice optimal. (Instance is negative only if all component classifiers mark it as negative.) xaxa x-x- wiwi wjwj wkwk wlwl c1(x)c1(x) xaxa x-x- wmwm wnwn wowo c2(x)c2(x) Example:

Evading Disjoint Conjunctions Theorem: By repeating linear attack with different constraints, we can efficiently find an attack that is at most twice optimal. Proof Sketch: Each component classifier has some optimal evasion. The optimal overall attack is the cheapest of these attacks. Running the linear attack once finds a good evasion against some classifier. Since it’s an evasion, one classifier must be negative. All feature changes for other classifiers can be removed. Since no individual or pair of changes reduces the cost, this evasion is at most twice optimal. By rerunning the linear attack restricted to features we haven’t used before, eventually we will find good evasions against all component classifiers. (Instance is negative if any component classifier marks it as negative.)

Experiments Data: 2005 TREC spam corpus Component classifiers: LR (SVM, NB in paper) Features partitioned into 3 or 5 sets: – Randomly – Spammy / Neutral / Hammy [Jorgensen et al., 2008] Fixed overall false negative rate to 10%. We attempted to disguise 100 different spams. To make this more challenging, we first added 100 random “spammy” features to each spam.

Results: Attack Cost

Results: Attack Optimality

Results: Attack Efficiency Number of queries before algorithms terminate: Conjunction: ~1,000,000 (Restricted: ~50,000) Disjunction: ~10,000,000 (Restricted: ~700,000)

1 million queries is not very efficient! The purpose of this experiment is to understand how performance depends on different factors, not the exact number of queries. In practice, the adversary’s job is much easier: – We added 100 spammy features to make it harder. – Additional background knowledge could make this much easier. – Restricted vocabulary reduces queries 10x with minimal increase in attack cost (90% of the time, still within 2x of optimal) – Attackers don’t need guarantees of optimality.

Results: Attack Efficiency Number of queries before our attack is within twice optimal: Conjunction: ~3,000 / ~100,000Disjunction: ~10,000 / ~300,000 Attacks are even easier with background knowledge and without 100 spammy words.

Discussion and Conclusion Evading discrete classifiers is provably harder than evading continuous classifiers. – Linear: k-approximation vs. (1+ε)-approximation – Polytope: Exponential vs. polynomial queries Interesting sub-classes of discrete non-linear classifiers are still vulnerable. – Disjoint features are a sufficient condition – Open question: What other sub-classes are vulnerable? Conjunction (convex spam) is theoretically harder but practically easier. – In addition to worst-case bounds, we need realistic simulations that can be applied to specific classifiers.