Regret Minimizing Audits: A Learning-theoretic Basis for Privacy Protection Jeremiah Blocki, Nicolas Christin, Anupam Datta, Arunesh Sinha Carnegie Mellon.

Slides:



Advertisements
Similar presentations
IDSIA Lugano Switzerland Master Algorithms for Active Experts Problems based on Increasing Loss Values Jan Poland and Marcus Hutter Defensive Universal.
Advertisements

Regret to the Best vs. Regret to the Average Eyal Even-Dar Computer and Information Science University of Pennsylvania Collaborators: Michael Kearns (Penn)
Policy Auditing over Incomplete Logs: Theory, Implementation and Applications Deepak Garg 1, Limin Jia 2 and Anupam Datta 2 1 MPI-SWS (work done at Carnegie.
Adaptive Regret Minimization in Bounded Memory Games Jeremiah Blocki, Nicolas Christin, Anupam Datta, Arunesh Sinha 1 GameSec 2013 – Invited Paper.
On-line learning and Boosting
Lecturer: Moni Naor Algorithmic Game Theory Uri Feige Robi Krauthgamer Moni Naor Lecture 8: Regret Minimization.
A Prior-Free Revenue Maximizing Auction for Secondary Spectrum Access Ajay Gopinathan and Zongpeng Li IEEE INFOCOM 2011, Shanghai, China.
DARWIN: Distributed and Adaptive Reputation Mechanism for Wireless Ad- hoc Networks CHEN Xiao Wei, Cheung Siu Ming CSE, CUHK May 15, 2008 This talk is.
Randomized Sensing in Adversarial Environments Andreas Krause Joint work with Daniel Golovin and Alex Roper International Joint Conference on Artificial.
LADIS workshop (Oct 11, 2009) A Case for the Accountable Cloud Andreas Haeberlen MPI-SWS.
A Simple Distribution- Free Approach to the Max k-Armed Bandit Problem Matthew Streeter and Stephen Smith Carnegie Mellon University.
Study Group Randomized Algorithms 21 st June 03. Topics Covered Game Tree Evaluation –its expected run time is better than the worst- case complexity.
Regret Minimization and the Price of Total Anarchy Paper by A. Blum, M. Hajiaghayi, K. Ligett, A.Roth Presented by Michael Wunder.
Online learning, minimizing regret, and combining expert advice
No-Regret Algorithms for Online Convex Programs Geoffrey J. Gordon Carnegie Mellon University Presented by Nicolas Chapados 21 February 2007.
Tuning bandit algorithms in stochastic environments The 18th International Conference on Algorithmic Learning Theory October 3, 2007, Sendai International.
1 Learning with continuous experts using Drifting Games work with Robert E. Schapire Princeton University work with Robert E. Schapire Princeton University.
R EGRET M INIMIZATION IN B OUNDED M EMORY G AMES Jeremiah Blocki Nicolas Christin Anupam Datta Arunesh Sinha Work in progress.
Questions?. Setting a reward function, with and without subgoals Difference between agent and environment AI for games, Roomba Markov Property – Broken.
Achieving Byzantine Agreement and Broadcast against Rational Adversaries Adam Groce Aishwarya Thiruvengadam Ateeq Sharfuddin CMSC 858F: Algorithmic Game.
©2006 OLC 1 Process Management: The Foundation for Achieving Organizational Excellence Process Management Implementation Worldwide.
FilterBoost: Regression and Classification on Large Datasets Joseph K. Bradley 1 and Robert E. Schapire 2 1 Carnegie Mellon University 2 Princeton University.
Sogang University ICC Lab Using Game Theory to Analyze Wireless Ad Hoc networks.
Regret Minimizing Audits: A Learning-theoretic Basis for Privacy Protection Jeremiah Blocki, Nicolas Christin, Anupam Datta, Arunesh Sinha Carnegie Mellon.
Algorithmic Complexity Nelson Padua-Perez Bill Pugh Department of Computer Science University of Maryland, College Park.
Mortal Multi-Armed Bandits Deepayan Chakrabarti,Yahoo! Research Ravi Kumar,Yahoo! Research Filip Radlinski, Microsoft Research Eli Upfal,Brown University.
The Value of Knowing a Demand Curve: Regret Bounds for Online Posted-Price Auctions Bobby Kleinberg and Tom Leighton.
Mechanism Design. Overview Incentives in teams (T. Groves (1973)) Algorithmic mechanism design (Nisan and Ronen (2000)) - Shortest Path - Task Scheduling.
AWESOME: A General Multiagent Learning Algorithm that Converges in Self- Play and Learns a Best Response Against Stationary Opponents Vincent Conitzer.
Shuchi Chawla, Carnegie Mellon University Static Optimality and Dynamic Search Optimality in Lists and Trees Avrim Blum Shuchi Chawla Adam Kalai 1/6/2002.
Dasgupta, Kalai & Monteleoni COLT 2005 Analysis of perceptron-based active learning Sanjoy Dasgupta, UCSD Adam Tauman Kalai, TTI-Chicago Claire Monteleoni,
Computing Sketches of Matrices Efficiently & (Privacy Preserving) Data Mining Petros Drineas Rensselaer Polytechnic Institute (joint.
Lecture 20: April 12 Introduction to Randomized Algorithms and the Probabilistic Method.
Anonymizing Web Services Through a Club Mechanism With Economic Incentives Mamata Jenamani Leszek Lilien Bharat Bhargava Department of Computer Sciences.
Handling Advertisements of Unknown Quality in Search Advertising Sandeep Pandey Christopher Olston (CMU and Yahoo! Research)
Learning and Planning for POMDPs Eyal Even-Dar, Tel-Aviv University Sham Kakade, University of Pennsylvania Yishay Mansour, Tel-Aviv University.
Experts Learning and The Minimax Theorem for Zero-Sum Games Maria Florina Balcan December 8th 2011.
Commitment without Regrets: Online Learning in Stackelberg Security Games Nika Haghtalab Carnegie Mellon University Joint work with Maria-Florina Balcan,
Online Learning Algorithms
MAKING COMPLEX DEClSlONS
Online Oblivious Routing Nikhil Bansal, Avrim Blum, Shuchi Chawla & Adam Meyerson Carnegie Mellon University 6/7/2003.
online convex optimization (with partial information)
General Polynomial Time Algorithm for Near-Optimal Reinforcement Learning Duke University Machine Learning Group Discussion Leader: Kai Ni June 17, 2005.
Advertisement 1. Audit Mechanisms for Provable Risk Management and Accountable Data Governance Jeremiah Blocki, Nicolas Christin, Anupam Datta, Arunesh.
Trust-Aware Optimal Crowdsourcing With Budget Constraint Xiangyang Liu 1, He He 2, and John S. Baras 1 1 Institute for Systems Research and Department.
A Learning-Based Approach to Reactive Security * Ben Rubinstein Microsoft Research Silicon Valley With: Adam Barth 1, Mukund Sundararajan 2, John Mitchell.
Class 5 Architecture-Based Self-Healing Systems David Garlan Carnegie Mellon University.
Regret Minimizing Equilibria of Games with Strict Type Uncertainty Stony Brook Conference on Game Theory Nathanaël Hyafil and Craig Boutilier Department.
1 Markov Decision Processes Infinite Horizon Problems Alan Fern * * Based in part on slides by Craig Boutilier and Daniel Weld.
Issues on the border of economics and computation נושאים בגבול כלכלה וחישוב Speaker: Dr. Michael Schapira Topic: Dynamics in Games (Part III) (Some slides.
Chapter 6 DECISION MAKING: THE ESSENCE OF THE MANAGER’S JOB 6.1 © 2003 Pearson Education Canada Inc.
Manipulating the Quota in Weighted Voting Games (M. Zuckerman, P. Faliszewski, Y. Bachrach, and E. Elkind) ‏ Presented by: Sen Li Software Technologies.
Information Theory for Mobile Ad-Hoc Networks (ITMANET): The FLoWS Project Competitive Scheduling in Wireless Networks with Correlated Channel State Ozan.
1 Monte-Carlo Planning: Policy Improvement Alan Fern.
Audit Games Jeremiah Blocki, Nicolas Christin, Anupam Datta, Ariel D. Procaccia, Arunesh Sinha 1 Carnegie Mellon University.
Dynamic Power Management Using Online Learning Gaurav Dhiman, Tajana Simunic Rosing (CSE-UCSD) Existing DPM policies do not adapt optimally with changing.
Bayesian Optimization. Problem Formulation Goal  Discover the X that maximizes Y  Global optimization Active experimentation  We can choose which values.
Online Learning Model. Motivation Many situations involve online repeated decision making in an uncertain environment. Deciding how to invest your money.
Carnegie Mellon University
Announcements Homework 1 Full assignment posted..
Game Theory Just last week:
Introduction to Randomized Algorithms and the Probabilistic Method
Rational Decision Making 8-step Process
Deploying PAWS: Field Optimization of the Protection Assistant for Wildlife Security Fang et al.
Distributed Submodular Maximization in Massive Datasets
Differential Privacy in Practice
Janardhan Rao (Jana) Doppa, Alan Fern, and Prasad Tadepalli
The
The Nonstochastic Multiarmed Bandit Problem
CASE − Cognitive Agents for Social Environments
Presentation transcript:

Regret Minimizing Audits: A Learning-theoretic Basis for Privacy Protection Jeremiah Blocki, Nicolas Christin, Anupam Datta, Arunesh Sinha Carnegie Mellon University

Motivation Goal: treatment Rigid access control hinders treatment Permissive access control ⇒ privacy violations Breach 2

A real problem 3

Audits Audits: one way to address the problem ◦ Permissive access control  If in doubt allow access ◦ Log the accesses ◦ Human auditors review the accesses later and find violations Adhoc approaches in practice ◦ FairWarning audit tool implements simple heuristics, e.g., flag all celebrity access 4

Desiderata Principled study of the audit process ◦ A model for audit process ◦ Properties of the audit mechanism ◦ Audit mechanism which provably satisfies the property 5

Auditing Challenges Organization’s economic tradeoff Employee’s incentives unknown How to optimally allocate budget for auditing, with no knowledge about adversary incentives? 6 Reputation loss Audit cost

Audit Algorithm by Example OverviewAudit ModelLow Regret Algorithm Auditing budget: $3000/ cycle Cost for one inspection: $ 1 00 Only 30 inspections per cycle Auditor 1 00 accesses 30 accesses 70 accesses Access divided into 2 types Loss from 1 violation (internal, external) $500, $ $250, $500

Audit Algorithm Choices 8 Only 30 inspections Consider 4 possible allocations of the available 30 inspections 1.0 Weights Choose allocation probabilistically based on weights OverviewAudit ModelLow Regret Algorithm

No. of Access Audit Algorithm Run Updated weights Observed Loss $2000$ 1 500$ $750$ $ Learn from experience: weights updated using observed and estimated loss 2 4 Actual Violation Ext. Caught Int. Caught OverviewAudit ModelLow Regret Algorithm Estimated Loss

Main Contributions  A game model for the audit process  Defining a desirable property of audit mechanisms, namely low regret  An efficient audit mechanism RMA that provably achieves low regret o Better bound on regret than existing algorithms that achieve low regret 10 OverviewAudit ModelLow Regret Algorithm

Repeated Game Model Game model The interaction repeats for each audit cycle (typically called rounds of repeated game) Typical actions in one round ◦ Emp action: (access, violate) = ([30,70], [2,4]) ◦ Org action: inspection = ([ 1 0,20]) Inspect Reputation loss Audit Cost Access, Violate One audit cycle (round) 11 Imperfection OverviewAudit ModelLow Regret Algorithm

Game Payoffs Organization’s payoff ◦ Audit cost depends on the number of inspections ◦ Reputation loss depends on the number of violations caught Employee’s payoff unknown Reputation loss Audit cost 12 OverviewAudit ModelLow Regret Algorithm

Regret Intuition 13 Is it possible to audit as well as the best strategy in hindsight ? OverviewAudit ModelLow Regret Algorithm

Regret by Example $5 $6 $0$ ,1 3, 2 Payoff of Org only Players Emp Org: s Round 1 3, 1 2 ( 1 ) Round 2 3,2 1 (-5) Total Payoff Unknown -4 Org : s 1 1 (2) 1 (-5) -3 Total regret(s, s 1 ) = (–5) – (–6) = 1 regret(s, s 1 ) = 1 /2 Strategy: outputs an action for every round Emp Org 14 Players Emp Org:s Round 1 3, 1 2 ( $6 ) Round 2 3, 2 1 ( $0) Total Payoff Unknown $6 Org:s 1 1 ($5) 1( $0) $5 OverviewAudit ModelLow Regret Algorithm

Meaning of Regret Low regret of s w.r.t. s 1 means s performs as well as s 1 Desirable property of an audit mechanism ◦ Low regret w.r.t all strategies in a given set of strategies ◦ regret → 0 as T → ∞ 15 OverviewAudit ModelLow Regret Algorithm

Regret minimization Multiplicative weight update (MWU) ◦ is a standard algorithm that achieves low regret w.r.t. to all strategies in a given set The regret bound of MWU is ◦ N: number of strategies in the given set ◦ T: number of rounds of the game ◦ All payoffs scaled to lie in [0, 1 ] 16 OverviewAudit ModelLow Regret Algorithm

Why not MWU? Imperfect information ◦ Org never learns the true action (violation) of the employee ◦ RMA regret bound: O((ln N)/T)  Best known bounds [ACFS03] : O((N 1 /3 ln N)/T 1 /3 )  Idea: estimate the payoff that would have been received Sleeping strategies: unavailable strategies ◦ Some inspections unavailable due to budgetary constraints ◦ We use techniques from [BM05] 17 [ACFS03] P. Auer, N. Cesa-Bianchi, Y. Freund, R. Schapire, “The nonstochastic multiarmed bandit problem,” SIAM Journal on Computing, 2003 [BM05] A. Blum and Y. Mansour, “From external to internal regret,”in COLT 2005 OverviewAudit ModelLow Regret Algorithm

Regret Minimizing Audits (RMA) 18 New audit cycle starts. Find AWAKE Pick s in AWAKE with probability D t (s) ∝ w s Update weight* of strategies s in AWAKE Estimate payoff vector Pay using Pay(s) Violation caught; obtain payoff Pay(s) w s = 1 for all strategies s * OverviewAudit ModelLow Regret Algorithm

Guarantees of RMA With probability RMA achieves the regret bound ◦ N is the set of strategies ◦ T is the number of rounds ◦ All payoffs scaled to lie in [0, 1 ] 19 OverviewAudit ModelLow Regret Algorithm

Related Work Authorization proof recorded in audit log [Vaughan et al. 2008] Analyze audit logs to detect and resolve access control policy misconfigurations [Bauer et al. 2008] Mechanically checkable complaince proof constructed using evidence from audit logs [Cederquist et al. 2007] Mechanically checking policy compliance over incomplete audit logs [Garg et al. 2011] 20

Take Away Message Future Work ◦ Evaluation over real hospital audit logs ◦ Analyze performance with more complex adversary models  Worst case + rational Learning technique for effective auditing with imperfect information

22