Foundations of Adversarial Learning Daniel Lowd, University of Washington Christopher Meek, Microsoft Research Pedro Domingos, University of Washington.

Slides:



Advertisements
Similar presentations
Statistical Machine Learning- The Basic Approach and Current Research Challenges Shai Ben-David CS497 February, 2007.
Advertisements

On-line learning and Boosting
Fast Algorithms For Hierarchical Range Histogram Constructions
Randomized Sensing in Adversarial Environments Andreas Krause Joint work with Daniel Golovin and Alex Roper International Joint Conference on Artificial.
Game Theoretical Insights in Strategic Patrolling: Model and Analysis Nicola Gatti – DEI, Politecnico di Milano, Piazza Leonardo.
Boosting Approach to ML
Partitioned Logistic Regression for Spam Filtering Ming-wei Chang University of Illinois at Urbana-Champaign Wen-tau Yih and Christopher Meek Microsoft.
On the Hardness of Evading Combinations of Linear Classifiers Daniel Lowd University of Oregon Joint work with David Stevens.
Efficient Query Evaluation on Probabilistic Databases
Models and Security Requirements for IDS. Overview The system and attack model Security requirements for IDS –Sensitivity –Detection Analysis methodology.
Assuming normally distributed data! Naïve Bayes Classifier.
Northwestern University Winter 2007 Machine Learning EECS Machine Learning Lecture 13: Computational Learning Theory.
Probably Approximately Correct Model (PAC)
Adversarial Learning: Practice and Theory Daniel Lowd University of Washington July 14th, 2006 Joint work with Chris Meek, Microsoft Research “If you know.
Foundations of Adversarial Learning Daniel Lowd, University of Washington Christopher Meek, Microsoft Research Pedro Domingos, University of Washington.
Bing LiuCS Department, UIC1 Learning from Positive and Unlabeled Examples Bing Liu Department of Computer Science University of Illinois at Chicago Joint.
Experts and Boosting Algorithms. Experts: Motivation Given a set of experts –No prior information –No consistent behavior –Goal: Predict as the best expert.
Online Learning Algorithms
Learning at Low False Positive Rate Scott Wen-tau Yih Joshua Goodman Learning for Messaging and Adversarial Problems Microsoft Research Geoff Hulten Microsoft.
Good Word Attacks on Statistical Spam Filters Daniel Lowd University of Washington (Joint work with Christopher Meek, Microsoft Research)
by B. Zadrozny and C. Elkan
1 Naïve Bayes Models for Probability Estimation Daniel Lowd University of Washington (Joint work with Pedro Domingos)
1 Logistic Regression Adapted from: Tom Mitchell’s Machine Learning Book Evan Wei Xiang and Qiang Yang.
Trust-Aware Optimal Crowdsourcing With Budget Constraint Xiangyang Liu 1, He He 2, and John S. Baras 1 1 Institute for Systems Research and Department.
Man vs. Machine: Adversarial Detection of Malicious Crowdsourcing Workers Gang Wang, Tianyi Wang, Haitao Zheng, Ben Y. Zhao, UC Santa Barbara, Usenix Security.
Bayesian Networks Martin Bachler MLA - VO
RESOURCES, TRADE-OFFS, AND LIMITATIONS Group 5 8/27/2014.
ECE 8443 – Pattern Recognition Objectives: Error Bounds Complexity Theory PAC Learning PAC Bound Margin Classifiers Resources: D.M.: Simplified PAC-Bayes.
Télécom 2A – Algo Complexity (1) Time Complexity and the divide and conquer strategy Or : how to measure algorithm run-time And : design efficient algorithms.
Partially Supervised Classification of Text Documents by Bing Liu, Philip Yu, and Xiaoli Li Presented by: Rick Knowles 7 April 2005.
Potential-Based Agnostic Boosting Varun Kanade Harvard University (joint work with Adam Tauman Kalai (Microsoft NE))
Major objective of this course is: Design and analysis of modern algorithms Different variants Accuracy Efficiency Comparing efficiencies Motivation thinking.
Online Algorithms By: Sean Keith. An online algorithm is an algorithm that receives its input over time, where knowledge of the entire input is not available.
Support Vector Machines Reading: Ben-Hur and Weston, “A User’s Guide to Support Vector Machines” (linked from class web page)
Empirical Research Methods in Computer Science Lecture 7 November 30, 2005 Noah Smith.
“Artificial Intelligence” in my research Seung-won Hwang Department of CSE POSTECH.
1 Bayesian Methods. 2 Naïve Bayes New data point to classify: X=(x 1,x 2,…x m ) Strategy: – Calculate P(C i /X) for each class C i. – Select C i for which.
Inference Complexity As Learning Bias Daniel Lowd Dept. of Computer and Information Science University of Oregon Joint work with Pedro Domingos.
Spam Detection Ethan Grefe December 13, 2013.
BLAST: Basic Local Alignment Search Tool Altschul et al. J. Mol Bio CS 466 Saurabh Sinha.
Slides for “Data Mining” by I. H. Witten and E. Frank.
Machine Learning Tom M. Mitchell Machine Learning Department Carnegie Mellon University Today: Computational Learning Theory Probably Approximately.
1/16 Seeing through M IST given a Small Fraction of an RSA Private Key Colin D. Walter Comodo Research Lab (Bradford, UK)
CS 8751 ML & KDDComputational Learning Theory1 Notions of interest: efficiency, accuracy, complexity Probably, Approximately Correct (PAC) Learning Agnostic.
Machine Learning in Practice Lecture 10 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
A Kernel Approach for Learning From Almost Orthogonal Pattern * CIS 525 Class Presentation Professor: Slobodan Vucetic Presenter: Yilian Qin * B. Scholkopf.
Classification using Co-Training
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
Machine Learning Chapter 7. Computational Learning Theory Tom M. Mitchell.
More on HMMs and Multiple Sequence Alignment BMI/CS 776 Mark Craven March 2002.
Data Mining CH6 Implementation: Real machine learning schemes(2) Reporter: H.C. Tsai.
On the Optimality of the Simple Bayesian Classifier under Zero-One Loss Pedro Domingos, Michael Pazzani Presented by Lu Ren Oct. 1, 2007.
Introduction to Information Retrieval Introduction to Information Retrieval Lecture 15: Text Classification & Naive Bayes 1.
Unconstrained Submodular Maximization Moran Feldman The Open University of Israel Based On Maximizing Non-monotone Submodular Functions. Uriel Feige, Vahab.
1 CS 391L: Machine Learning: Computational Learning Theory Raymond J. Mooney University of Texas at Austin.
Advanced Algorithms Analysis and Design
Boosted Augmented Naive Bayes. Efficient discriminative learning of
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Introduction to Machine Learning
Alternative Representations for Artificial Immune Systems
KDD 2004: Adversarial Classification
Extensive-form games and how to solve them
Objective of This Course
CSCI B609: “Foundations of Data Science”
Computational Learning Theory
Computational Learning Theory
Chapter 11 Limitations of Algorithm Power
What is The Optimal Number of Features
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Normal Form (Matrix) Games
Presentation transcript:

Foundations of Adversarial Learning Daniel Lowd, University of Washington Christopher Meek, Microsoft Research Pedro Domingos, University of Washington

Motivation Many adversarial problems  Spam filtering  Intrusion detection  Malware detection  New ones every year! Want general-purpose solutions We can gain much insight by modeling adversarial situations mathematically

Outline Problem definitions Anticipating adversaries (Dalvi et al., 2004)  Goal: Defeat adaptive adversary  Assume: Perfect information, optimal short-term strategies  Results: Vastly better classifier accuracy Reverse engineering classifiers (Lowd & Meek, 2005a,b)  Goal: Assess classifier vulnerability  Assume: Membership queries from adversary  Results: Theoretical bounds, practical attacks Conclusion

Definitions X1X1 X2X2 x X1X1 X2X2 x + - X1X1 X2X2 Instance space Classifier Adversarial cost function c(x): X  {+,  } c  C, concept class (e.g., linear classifier) X = {X 1, X 2, …, X n } Each X i is a feature Instances, x  X (e.g., s) a(x): X  R a  A (e.g., more legible spam is better)

Adversarial scenario Classifier’s Task: Choose new c’(x) minimize (cost-sensitive) error Adversary’s Task: Choose x to minimize a(x) subject to c(x) = 

This is a game! Adversary’s actions: {x  X} Classifier’s actions: {c  C} Assume perfect information A Nash equilibrium exists… …but finding it is triply exponential (in easy cases).

Tractable approach Start with a trained classifier  Use cost-sensitive naïve Bayes  Assume: training data is untainted Compute adversary’s best action, x  Use cost: a(x) = Σ i w(x i, b i )  Solve knapsack-like problem with dynamic programming  Assume: that the classifier will not modify c(x) Compute classifier’s optimal response, c’(x)  For given x, compute probability it was modified by adversary  Assume: the adversary is using the optimal strategy By anticipating the adversary’s strategy, we can defeat it!

Evaluation: spam Data: -Data Scenarios  Plain (PL)  Add Words (AW)  Synonyms (SYN)  Add Length (AL) Similar results with Ling-Spam, different classifier costs Score

Outline Problem definitions Anticipating adversaries (Dalvi et al., 2004)  Goal: Defeat adaptive adversary  Assume: Perfect information, optimal short-term strategies  Results: Vastly better classifier accuracy Reverse engineering classifiers (Lowd & Meek, 2005a,b)  Goal: Assess classifier vulnerability  Assume: Membership queries from adversary  Results: Theoretical bounds, practical attacks Conclusion

Imperfect information What can an adversary accomplish with limited knowledge of the classifier? Goals:  Understand classifier’s vulnerabilities  Understand our adversary’s likely strategies “If you know the enemy and know yourself, you need not fear the result of a hundred battles.” -- Sun Tzu, 500 BC

Adversarial Classification Reverse Engineering (ACRE) + - Adversary’s Task: Minimize a(x) subject to c(x) =  Problem: The adversary doesn’t know c(x)!

Adversarial Classification Reverse Engineering (ACRE) Task: Minimize a(x) subject to c(x) =  Given: X1X1 X2X2 ?? ? ? ? ? ? ? - + –Full knowledge of a(x) –One positive and one negative instance, x + and x  –A polynomial number of membership queries Within a factor of k

Comparison to other theoretical learning methods Probably Approximately Correct (PAC): accuracy over same distribution Membership queries: exact classifier ACRE: single low-cost, negative instance

ACRE example X1X1 X2X2 X1X1 X2X2 xaxa Linear classifier: c(x) = +, iff (w  x > T) Linear cost function:

Linear classifiers with continuous features ACRE learnable within a factor of (1+  ) under linear cost functions Proof sketch  Only need to change the highest weight/cost feature  We can efficiently find this feature using line searches in each dimension X1X1 X2X2 xaxa

Linear classifiers with Boolean features Harder problem: can’t do line searches ACRE learnable within a factor of 2 if adversary has unit cost per change: xaxa x-x- wiwi wjwj wkwk wlwl wmwm c(x)c(x)

Algorithm Iteratively reduce the cost in two ways: 1. Remove any unnecessary change: O(n) 2. Replace any two changes with one: O(n 3 ) xaxa y wiwi wjwj wkwk wlwl c(x)c(x) wmwm x-x- xaxa y’ wiwi wjwj wkwk wlwl c(x)c(x) wpwp

Evaluation Classifiers: Naïve Bayes (NB), Maxent (ME) Data: 500k Hotmail messages, 250k features Adversary feature sets:  23,000 words (Dict)  1,000 random words (Rand) CostQueries Dict NB23261,000 Dict ME10119,000 Rand NB3123,000 Rand ME129,000

Finding features We can find good features (words) instead of good instances ( s) Passive attack: choose words common in English but uncommon in spam First-N attack: choose words that turn a “barely spam” into a non-spam Best-N attack: use “spammy” words to sort good words

Results Attack typeNaïve Bayes words (queries) Maxent words (queries) Passive 112 (0) 149 (0) First-N 59 (3,100) 20 (4,300) Best-N 29 (62,000) 9(69,000) ACRE (Rand) 31* (23,000) 12* (9,000) * words added + words removed

Conclusion Mathematical modeling is a powerful tool in adversarial situations  Game theory lets us make classifiers aware of and resistant to adversaries  Complexity arguments let us explore the vulnerabilities of our own systems This is only the beginning…  Can we weaken our assumptions?  Can we expand our scenarios?

Proof sketch (Contradiction) xaxa y wiwi wjwj wkwk wlwl c(x)c(x) wmwm x wpwp wrwr x’s average change is twice as good as y’s We can replace y’s two worst changes with x’s single best change But we already tried every such replacement! Suppose there is some negative instance x with less than half the cost of y: