Regret Minimizing Audits: A Learning-theoretic Basis for Privacy Protection Jeremiah Blocki, Nicolas Christin, Anupam Datta, Arunesh Sinha Carnegie Mellon.

Slides:



Advertisements
Similar presentations
Policy Auditing over Incomplete Logs: Theory, Implementation and Applications Deepak Garg 1, Limin Jia 2 and Anupam Datta 2 1 MPI-SWS (work done at Carnegie.
Advertisements

QoS-based Management of Multiple Shared Resources in Dynamic Real-Time Systems Klaus Ecker, Frank Drews School of EECS, Ohio University, Athens, OH {ecker,
Adaptive Regret Minimization in Bounded Memory Games Jeremiah Blocki, Nicolas Christin, Anupam Datta, Arunesh Sinha 1 GameSec 2013 – Invited Paper.
TAU Agent Team: Yishay Mansour Mariano Schain Tel Aviv University TAC-AA 2010.
Lecturer: Moni Naor Algorithmic Game Theory Uri Feige Robi Krauthgamer Moni Naor Lecture 8: Regret Minimization.
A Prior-Free Revenue Maximizing Auction for Secondary Spectrum Access Ajay Gopinathan and Zongpeng Li IEEE INFOCOM 2011, Shanghai, China.
DARWIN: Distributed and Adaptive Reputation Mechanism for Wireless Ad- hoc Networks CHEN Xiao Wei, Cheung Siu Ming CSE, CUHK May 15, 2008 This talk is.
Delay bounded Routing in Vehicular Ad-hoc Networks Antonios Skordylis Niki Trigoni MobiHoc 2008 Slides by Alex Papadimitriou.
A Simple Distribution- Free Approach to the Max k-Armed Bandit Problem Matthew Streeter and Stephen Smith Carnegie Mellon University.
Study Group Randomized Algorithms 21 st June 03. Topics Covered Game Tree Evaluation –its expected run time is better than the worst- case complexity.
Game Theory 1. Game Theory and Mechanism Design Game theory to analyze strategic behavior: Given a strategic environment (a “game”), and an assumption.
Markov Game Analysis for Attack and Defense of Power Networks Chris Y. T. Ma, David K. Y. Yau, Xin Lou, and Nageswara S. V. Rao.
Regret Minimizing Audits: A Learning-theoretic Basis for Privacy Protection Jeremiah Blocki, Nicolas Christin, Anupam Datta, Arunesh Sinha Carnegie Mellon.
Regret Minimization and the Price of Total Anarchy Paper by A. Blum, M. Hajiaghayi, K. Ligett, A.Roth Presented by Michael Wunder.
Online learning, minimizing regret, and combining expert advice
No-Regret Algorithms for Online Convex Programs Geoffrey J. Gordon Carnegie Mellon University Presented by Nicolas Chapados 21 February 2007.
1 Regret-based Incremental Partial Revelation Mechanism Design Nathanaël Hyafil, Craig Boutilier AAAI 2006 Department of Computer Science University of.
R EGRET M INIMIZATION IN B OUNDED M EMORY G AMES Jeremiah Blocki Nicolas Christin Anupam Datta Arunesh Sinha Work in progress.
Questions?. Setting a reward function, with and without subgoals Difference between agent and environment AI for games, Roomba Markov Property – Broken.
Achieving Byzantine Agreement and Broadcast against Rational Adversaries Adam Groce Aishwarya Thiruvengadam Ateeq Sharfuddin CMSC 858F: Algorithmic Game.
©2006 OLC 1 Process Management: The Foundation for Achieving Organizational Excellence Process Management Implementation Worldwide.
Ýmir Vigfússon IBM Research Haifa Labs Ken Birman Cornell University Qi Huang Cornell University Deepak Nataraj Cornell University.
FilterBoost: Regression and Classification on Large Datasets Joseph K. Bradley 1 and Robert E. Schapire 2 1 Carnegie Mellon University 2 Princeton University.
Sogang University ICC Lab Using Game Theory to Analyze Wireless Ad Hoc networks.
Fast Iterative Alignment of Pose Graphs with Poor Initial Estimates Edwin Olson John Leonard, Seth Teller
Algorithmic Complexity Nelson Padua-Perez Bill Pugh Department of Computer Science University of Maryland, College Park.
The Value of Knowing a Demand Curve: Regret Bounds for Online Posted-Price Auctions Bobby Kleinberg and Tom Leighton.
Mechanism Design. Overview Incentives in teams (T. Groves (1973)) Algorithmic mechanism design (Nisan and Ronen (2000)) - Shortest Path - Task Scheduling.
Agent Technology for e-Commerce Chapter 10: Mechanism Design Maria Fasli
AWESOME: A General Multiagent Learning Algorithm that Converges in Self- Play and Learns a Best Response Against Stationary Opponents Vincent Conitzer.
SECOND PART: Algorithmic Mechanism Design. Mechanism Design MD is a subfield of economic theory It has a engineering perspective Designs economic mechanisms.
Robust Mechanisms for Information Elicitation Aviv Zohar & Jeffrey S. Rosenschein The Hebrew University.
Trust Based Mechanism Design. Use MD Motivation Fuse the fields of trust-modelling and mechanism design Trust measures how good an interaction partner.
Computing Sketches of Matrices Efficiently & (Privacy Preserving) Data Mining Petros Drineas Rensselaer Polytechnic Institute (joint.
Lecture 20: April 12 Introduction to Randomized Algorithms and the Probabilistic Method.
Anonymizing Web Services Through a Club Mechanism With Economic Incentives Mamata Jenamani Leszek Lilien Bharat Bhargava Department of Computer Sciences.
Experts Learning and The Minimax Theorem for Zero-Sum Games Maria Florina Balcan December 8th 2011.
Commitment without Regrets: Online Learning in Stackelberg Security Games Nika Haghtalab Carnegie Mellon University Joint work with Maria-Florina Balcan,
Online Learning Algorithms
01/16/2002 Reliable Query Reporting Project Participants: Rajgopal Kannan S. S. Iyengar Sudipta Sarangi Y. Rachakonda (Graduate Student) Sensor Networking.
New Protocols for Remote File Synchronization Based on Erasure Codes Utku Irmak Svilen Mihaylov Torsten Suel Polytechnic University.
MAKING COMPLEX DEClSlONS
online convex optimization (with partial information)
CPS 173 Mechanism design Vincent Conitzer
Advertisement 1. Audit Mechanisms for Provable Risk Management and Accountable Data Governance Jeremiah Blocki, Nicolas Christin, Anupam Datta, Arunesh.
Trust-Aware Optimal Crowdsourcing With Budget Constraint Xiangyang Liu 1, He He 2, and John S. Baras 1 1 Institute for Systems Research and Department.
A Learning-Based Approach to Reactive Security * Ben Rubinstein Microsoft Research Silicon Valley With: Adam Barth 1, Mukund Sundararajan 2, John Mitchell.
Mechanism design. Goal of mechanism design Implementing a social choice function f(u 1, …, u |A| ) using a game Center = “auctioneer” does not know the.
Regret Minimizing Equilibria of Games with Strict Type Uncertainty Stony Brook Conference on Game Theory Nathanaël Hyafil and Craig Boutilier Department.
Linear Programming Erasmus Mobility Program (24Apr2012) Pollack Mihály Engineering Faculty (PMMK) University of Pécs João Miranda
Issues on the border of economics and computation נושאים בגבול כלכלה וחישוב Speaker: Dr. Michael Schapira Topic: Dynamics in Games (Part III) (Some slides.
Adversarial Search Chapter Games vs. search problems "Unpredictable" opponent  specifying a move for every possible opponent reply Time limits.
Information Theory for Mobile Ad-Hoc Networks (ITMANET): The FLoWS Project Competitive Scheduling in Wireless Networks with Correlated Channel State Ozan.
Audit Games Jeremiah Blocki, Nicolas Christin, Anupam Datta, Ariel D. Procaccia, Arunesh Sinha 1 Carnegie Mellon University.
Dynamic Power Management Using Online Learning Gaurav Dhiman, Tajana Simunic Rosing (CSE-UCSD) Existing DPM policies do not adapt optimally with changing.
Principles of Management
Online Learning Model. Motivation Many situations involve online repeated decision making in an uncertain environment. Deciding how to invest your money.
Artificial Intelligence AIMA §5: Adversarial Search
Announcements Homework 1 Full assignment posted..
Game Theory Just last week:
Better Algorithms for Better Computers
Task: It is necessary to choose the most suitable variant from some set of objects by those or other criteria.
Communication Complexity as a Lower Bound for Learning in Games
Prof. Marie desJardins September 20, 2010
Deploying PAWS: Field Optimization of the Protection Assistant for Wildlife Security Fang et al.
Janardhan Rao (Jana) Doppa, Alan Fern, and Prasad Tadepalli
Convergence, Targeted Optimality, and Safety in Multiagent Learning
Multiagent Systems Repeated Games © Manfred Huber 2018.
Minimax strategies, alpha beta pruning
Dr. Arslan Ornek MATHEMATICAL MODELS
Presentation transcript:

Regret Minimizing Audits: A Learning-theoretic Basis for Privacy Protection Jeremiah Blocki, Nicolas Christin, Anupam Datta, Arunesh Sinha Carnegie Mellon University

Motivation Goal: treatment Rigid access control hinders treatment Permissive access control ⇒ human in the loop ⇒ privacy violations Breach 2

A real problem 3

Enforcement Using Audits Permissive access control ◦ If in doubt allow access Log the accesses Review the accesses later and find violations Adhoc approaches in practice ◦ FairWarning audit tool implements simple heuristics, e.g., flag all celebrity access 4

Humans are the weakest link 5 Design an audit mechanism robust to irresponsible/careless human behavior Audit based enforcement involves human participation ◦ Employees trusted to act responsibly

Desiderata Principled study of the audit process ◦ A model (including humans) for audit process ◦ Properties of the audit mechanism ◦ Audit mechanism which provably satisfies the property 6

Audit Algorithm by Example OverviewAudit ModelLow Regret Algorithm Auditing budget: $3000/ cycle Cost for one inspection: $ 1 00 Only 30 inspections per cycle Auditor 1 00 accesses 30 accesses 70 accesses Access divided into 2 types Reputation Loss from 1 violation (internal, external) $500, $ $250, $500

Audit Algorithm Choices 8 Only 30 inspections Employee incentive unknown Consider 4 possible allocations of the available 30 inspections 1.0 Weights Choose allocation probabilistically based on weights OverviewAudit ModelLow Regret Algorithm

No. of Access Audit Algorithm Run Updated weights Observed Loss $2000$ 1 500$ $750$ $ Learn from experience: weights updated using observed and estimated loss 2 4 Actual Violation Ext. Caught Int. Caught OverviewAudit ModelLow Regret Algorithm Estimated Loss

Main Contributions  A game model for the audit process  Defining a desirable property of audit mechanisms, namely low regret  An efficient audit mechanism RMA that provably achieves low regret o Better bound on regret than existing algorithms that achieve low regret 10 OverviewAudit ModelLow Regret Algorithm

Repeated Game Model Game model Typical actions in one round ◦ Emp action: (access, violate) = ([30,70], [2,4]) ◦ Org action: inspection = ([ 1 0,20]) Inspect Reputation loss Audit Cost Access, Violate One audit cycle (round) 11 Imperfection OverviewAudit ModelLow Regret Algorithm

Game Payoffs Organization’s payoff ◦ Audit cost depends on the number of inspections ◦ Reputation loss depends on the number of violations caught Employee’s payoff unknown ◦ Guarantees of any audit mechanism in this model holds irrespective of behavior of the employee Reputation loss Audit cost 12 OverviewAudit ModelLow Regret Algorithm

Regret by Example $5 $6 $0$ ,1 3, 2 Payoff of Org only Players Emp Org: s Round 1 3, 1 2 ( 1 ) Round 2 3,2 1 (-5) Total Payoff Unknown -4 Org : s 1 1 (2) 1 (-5) -3 Total regret(s, s 1 ) = (–5) – (–6) = 1 regret(s, s 1 ) = 1 /2 Strategy: outputs an action for every round Emp Org 13 Players Emp Org:s Round 1 3, 1 2 ( $6 ) Round 2 3, 2 1 ( $0) Total Payoff Unknown $6 Org:s 1 1 ($5) 1( $0) $5 OverviewAudit ModelLow Regret Algorithm

Meaning of Regret Low regret of s w.r.t. s 1 means s performs as well as s 1 Desirable property of an audit mechanism ◦ Low regret w.r.t all strategies in a given set of strategies ◦ regret → 0 as T → ∞ 14 OverviewAudit ModelLow Regret Algorithm

Regret minimization Multiplicative weight update (MWU) ◦ standard algorithm that achieves low regret w.r.t. to all strategies in a given set The regret bound of MWU is  N: number of strategies in the given set  T: number of rounds of the game  All payoffs scaled to lie in [0, 1 ] Why not MWU? ◦ Imperfect information, unavailable strategies 15 OverviewAudit ModelLow Regret Algorithm

Regret Minimizing Audits (RMA) 16 New audit cycle starts. Find AWAKE Pick s in AWAKE with probability D t (s) ∝ w s Update weight* of strategies s in AWAKE Estimate payoff vector Pay using Pay(s) Violation caught; obtain payoff Pay(s) w s = 1 for all strategies s * OverviewAudit ModelLow Regret Algorithm

Guarantees of RMA With probability RMA achieves the regret bound ◦ N is the set of strategies ◦ T is the number of rounds ◦ All payoffs scaled to lie in [0, 1 ] Better bound than any existing algorithm 17 OverviewAudit ModelLow Regret Algorithm

Take Away Message Future Work ◦ Evaluation over real hospital audit logs ◦ Analyze performance with more complex adversary models  Worst case + rational Optimize costs, given imperfect nature of periodic audits in a setting with adaptive adversaries (employees) whose incentives are not known. 18

19