Information Technology Michael Brand Joint work with David L. Dowe 8 February, 2016 Information Technology.

Slides:



Advertisements
Similar presentations
An Introduction to Game Theory Part V: Extensive Games with Perfect Information Bernhard Nebel.
Advertisements

Foundations of Cryptography Lecture 10 Lecturer: Moni Naor.
Congestion Games with Player- Specific Payoff Functions Igal Milchtaich, Department of Mathematics, The Hebrew University of Jerusalem, 1993 Presentation.
Chapter 6 Game Theory © 2006 Thomson Learning/South-Western.
EC3224 Autumn Lecture #04 Mixed-Strategy Equilibrium
© 2015 McGraw-Hill Education. All rights reserved. Chapter 15 Game Theory.
EC941 - Game Theory Lecture 7 Prof. Francesco Squintani
Lecture 16: Relativization Umans Complexity Theory Lecturess.
CSCI 4325 / 6339 Theory of Computation Zhixiang Chen Department of Computer Science University of Texas-Pan American.
Chapter 6 © 2006 Thomson Learning/South-Western Game Theory.
1 Languages. 2 A language is a set of strings String: A sequence of letters Examples: “cat”, “dog”, “house”, … Defined over an alphabet: Languages.
Game-Theoretic Approaches to Multi-Agent Systems Bernhard Nebel.
Complexity 11-1 Complexity Andrei Bulatov Space Complexity.
1 Introduction to Computability Theory Lecture13: Mapping Reductions Prof. Amos Israeli.
Courtesy Costas Busch - RPI1 A Universal Turing Machine.
CS5371 Theory of Computation
Lecture 9 Recursive and r.e. language classes
1 Module 9 Recursive and r.e. language classes –representing solvable and half-solvable problems Proofs of closure properties –for the set of recursive.
Lecture 8 Recursively enumerable (r.e.) languages
Programming the TM qa  (,q) (,q) q1q1 0q1q1 R q1q1 1q1q1 R q1q1  h  Qa  (,q) (,q) q1q1 0q2q2  q1q1 1q3q3  q1q1  h  q2q2 0q4q4 R q2q2 1q4q4.
AWESOME: A General Multiagent Learning Algorithm that Converges in Self- Play and Learns a Best Response Against Stationary Opponents Vincent Conitzer.
1 Languages and Finite Automata or how to talk to machines...
CHAPTER 4 Decidability Contents Decidable Languages
RELATIVIZATION CSE860 Vaishali Athale. Overview Introduction Idea behind “Relativization” Concept of “Oracle” Review of Diagonalization Proof Limits of.
Fall 2004COMP 3351 Reducibility. Fall 2004COMP 3352 Problem is reduced to problem If we can solve problem then we can solve problem.
1 Lecture 18 Closure Properties of Language class LFSA –Remember ideas used in solvable languages unit –Set complement –Set intersection, union, difference,
UNIT II: The Basic Theory Zero-sum Games Nonzero-sum Games Nash Equilibrium: Properties and Problems Bargaining Games Bargaining and Negotiation Review.
Introduction to Game Theory and Behavior Networked Life CIS 112 Spring 2009 Prof. Michael Kearns.
1 Decidability continued. 2 Undecidable Problems Halting Problem: Does machine halt on input ? State-entry Problem: Does machine enter state halt on input.
Automata & Formal Languages, Feodor F. Dragan, Kent State University 1 CHAPTER 5 Reducibility Contents Undecidable Problems from Language Theory.
Fall 2004COMP 3351 A Universal Turing Machine. Fall 2004COMP 3352 Turing Machines are “hardwired” they execute only one program A limitation of Turing.
On Bounded Rationality and Computational Complexity Christos Papadimitriou and Mihallis Yannakakis.
January 28, 2015CS21 Lecture 101 CS21 Decidability and Tractability Lecture 10 January 28, 2015.
Regular Expressions and Automata Chapter 2. Regular Expressions Standard notation for characterizing text sequences Used in all kinds of text processing.
Prof. Busch - LSU1 Turing Machines. Prof. Busch - LSU2 The Language Hierarchy Regular Languages Context-Free Languages ? ?
1 Module 10 Recursive and r.e. language classes –representing solvable and half-solvable problems Proofs of closure properties –for the set of recursive.
UNIT II: The Basic Theory Zero-sum Games Nonzero-sum Games Nash Equilibrium: Properties and Problems Bargaining Games Bargaining and Negotiation Review.
The Multiplicative Weights Update Method Based on Arora, Hazan & Kale (2005) Mashor Housh Oded Cats Advanced simulation methods Prof. Rubinstein.
Peter van Emde Boas: Games and Computer Science 1999 GAMES AND COMPUTER SCIENCE Theoretical Models 1999 Peter van Emde Boas References available at:
Presenter: Chih-Yuan Chou GA-BASED ALGORITHMS FOR FINDING EQUILIBRIUM 1.
1 1 CDT314 FABER Formal Languages, Automata and Models of Computation Lecture 15-1 Mälardalen University 2012.
. CLASSES RP AND ZPP By: SARIKA PAMMI. CONTENTS:  INTRODUCTION  RP  FACTS ABOUT RP  MONTE CARLO ALGORITHM  CO-RP  ZPP  FACTS ABOUT ZPP  RELATION.
MA/CSSE 474 Theory of Computation Enumerability Reduction.
MA/CSSE 474 Theory of Computation Decision Problems DFSMs.
A Universal Turing Machine
1 What is Game Theory About? r Analysis of situations where conflict of interests is present r Goal is to prescribe how conflicts can be resolved 2 2 r.
1 Linear Bounded Automata LBAs. 2 Linear Bounded Automata (LBAs) are the same as Turing Machines with one difference: The input string tape space is the.
1 Turing’s Thesis. 2 Turing’s thesis: Any computation carried out by mechanical means can be performed by a Turing Machine (1930)
Chap2. Language Acquisition: The Problem of Inductive Inference (2.1 ~ 2.2) Min Su Lee The Computational Nature of Language Learning and Evolution.
Zero-sum Games The Essentials of a Game Extensive Game Matrix Game Dominant Strategies Prudent Strategies Solving the Zero-sum Game The Minimax Theorem.
Lecture 24UofH - COSC Dr. Verma 1 COSC 3340: Introduction to Theory of Computation University of Houston Dr. Verma Lecture 24.
1 Course Overview Why this course “formal languages and automata theory?” What do computers really do? What are the practical benefits/application of formal.
1 Introduction to Turing Machines
Turing Machines Sections 17.6 – The Universal Turing Machine Problem: All our machines so far are hardwired. ENIAC
Recall last lecture and Nondeterministic TMs Ola Svensson.
1 Turing Machines. 2 The Language Hierarchy Regular Languages Context-Free Languages ? ?
MA/CSSE 474 Theory of Computation Decision Problems, Continued DFSMs.
Recursively Enumerable and Recursive Languages. Definition: A language is recursively enumerable if some Turing machine accepts it.
1 A Universal Turing Machine. 2 Turing Machines are “hardwired” they execute only one program A limitation of Turing Machines: Real Computers are re-programmable.
The Acceptance Problem for TMs
A Universal Turing Machine
Languages.
Busch Complexity Lectures: Turing Machines
CS154, Lecture 7: Turing Machines.
Communication Complexity as a Lower Bound for Learning in Games
LIMITS OF ALGORITHMIC COMPUTATION
Turing Machines Acceptors; Enumerators
Chapter Nine: Advanced Topics in Regular Languages
Formal Languages, Automata and Models of Computation
Recall last lecture and Nondeterministic TMs
Presentation transcript:

Information Technology Michael Brand Joint work with David L. Dowe 8 February, 2016 Information Technology

8 February, 2016The IMP game2 Three questions  Approximability –How much can (well chosen) elements from one set be made to resemble (arbitrary) elements from another set? –We consider languages and  Learning –How well can one predict a sequence by seeing its past elements?  Adversarial learning –Two adversaries try to predict each other’s moves and capitalise on the predictions. How well can each do? –Very hot topic, currently: Online bidding strategies. Poisoning attacks.

8 February, 2016The IMP game3 Major results

8 February, 2016The IMP game4 Major results (cntd) Informally:  Learning –Turing machines can learn by example beyond what is computable. –In fact, they can learn all R.E. and all co-R.E. languages (and more).  Adversarial learning –In an iterated game of matching pennies (a.k.a. “odds and evens”), the player choosing “evens” has a decisive advantage.

8 February, 2016The IMP game5 Caveat  Conclusions inevitably depend on one’s base definitions. –For approximability, for example, we used the DisSim metric, but other distance metrics would have yielded potentially different results.  The same goes for our definition of “to learn” that underpins the “learning” and “adversarial learning” results.  The literature has many definitions of “learnability”: –Solomonoff –E. M. Gold –Statistical Consistency –PAC –etc.  Our definition is not identical to any of these, but has a resemblance to all of them.

8 February, 2016The IMP game6 Our justifications  We give a single, unified framework within which all three problems (approximability, learnability, adversarial learning) can be investigated.  We want to explore the “game” aspects of adversarial learning, so naturally integrate tools from game theory (e.g., mixed strategies, Nash equilibria). –We begin by analysing adversarial learning, then take the other cases as special cases. –Traditional approaches typically begin with “learning”, and need special provisions for adversarial learning, sometimes losing entirely the “game” character and reducing the process to a one- player game. –We believe that our approach, retaining the “game” elements, is more natural.  The results are interesting!

8 February, 2016The IMP game7 The IMP set-up

8 February, 2016The IMP game8 A game of matching pennies Player “=“Player “≠“ Accept/ Reject

8 February, 2016The IMP game9 An iterated game of matching pennies Player “=“Player “≠“ Accept/ Reject Agent Inspector Final payoffs?

8 February, 2016The IMP game10 Why the strange payoffs?

8 February, 2016The IMP game11 An iterated game of matching pennies Player “=“Player “≠“ Accept/ Reject Agent Inspector Strategy Mixed strategy = distribution

8 February, 2016The IMP game12 Accept/ Reject Agent L=L= L≠L≠

8 February, 2016The IMP game13 Player “=“ strategyPlayer “≠“ strategy Accept/ Reject Agent L=L= L≠L≠ D=D= D≠D≠

8 February, 2016The IMP game14 Player “=“ strategyPlayer “≠“ strategy Accept/ Reject Agent Oracle. Does not need to be computable. Performs a xor over accept/reject choices. Inspector Observation: The key to enabling the learning from examples of incomputable functions is to have a method to generate the examples... L=L= L≠L≠ D=D= D≠D≠

8 February, 2016The IMP game15 Player “=“ strategyPlayer “≠“ strategy Accept/ Reject Agent Inspector Agents are effectively “restarted” at every iteration. The feedback from the inspector is their input string. L=L= L≠L≠ D=D= D≠D≠

8 February, 2016The IMP game16 Reminder of some (standard) definitions we’ll use

8 February, 2016The IMP game17 The Arithmetical Hierarchy

8 February, 2016The IMP game18 Nash equilibrium

8 February, 2016The IMP game19 Warm-up: halting Turing machines

8 February, 2016The IMP game20 Characterisation of Nash equilibria

8 February, 2016The IMP game21 –Will make at most X errors w.p. 1-ε, so maxmin=0. –Note: L 0,...,L X can be finitely encoded by (finite) T 0,...,T X. –Symmetrically, for any D =, define L ≠ to prove minmax=1. –Because maxmin≠minmax, no Nash equilibrium exists. Only change needed.

8 February, 2016The IMP game22 The general case

8 February, 2016The IMP game23 Adversarial learnability Can only lose a finite number of rounds against any agent!

8 February, 2016The IMP game24 Adversarial learnability (cntd)

8 February, 2016The IMP game25 Conventional learning

8 February, 2016The IMP game26 Nonadaptive strategies

8 February, 2016The IMP game27 Conventional learnability

8 February, 2016The IMP game28 Proof (general idea)

8 February, 2016The IMP game29 Proof (cntd)

8 February, 2016The IMP game30 Proof (cntd)  Complication #4: After the first guess, all remaining t-1 guesses happen at different rounds among the different agents. How can we ensure a 25% success rate for each guess? –Solution: We make sure all guesses are synchronised between agents. The way to do this is to pre-allocate for each of the t guesses an infinite sequence of rounds, such that in total these rounds amount to a density of 0 among all rounds. Each guess retains its pre-allocated rounds until it is falsified. Guesses all happen in pre-allocated places within these pre-allocated rounds. –The remaining rounds (forming the overwhelming majority) are used by the current “best guess”: the lowest-numbered hypothesis yet to be falsified. –Total success rate: t, for a sup of 1, as required.

8 February, 2016The IMP game31 Proof (cntd)  Complication #5: But we don’t know which co-R.E. function to emulate... –Solution: Instead of having t hypotheses, we have an infinite number of hypotheses, t for each co-R.E. function. We enumerate over all. –We pre-allocate an infinite number of bits to each of these infinite hypotheses, while still maintaining that their total density is 0.  Notably, if our learner was probabilistic, there was no need for a mixed strategy. –Although this, too, has its own complications...  However, we are able to prove that no pure-strategy deterministic agent can learn the co-R.E. languages.  This is a case where stochastic TMs have a provable advantage.

8 February, 2016The IMP game32 Approximation

8 February, 2016The IMP game33

8 February, 2016The IMP game34

8 February, 2016The IMP game35 Proof of lim inf claim  triangle(x) := 0,0,1,0,1,2,0,1,2,3,0,1,2,3,4,0,1,2,3,4,5,...  caf(x) := maximum y s.t. y!≤x.  Define L = by  L = emulates each language an infinite number of times.  Each time, it does so for a length that becomes an increasing proportion (with a lim of 1) of the total number of rounds so far.  Consider the subsequence relating to the correct guess for L ≠. This gives the lim inf result.

8 February, 2016The IMP game36 Proof of Theorem 1

8 February, 2016The IMP game37 Some open questions

QUESTIONS? Thank you! 8 February, 2016The IMP game38