Lecture VI: Adaptive Systems

Slides:



Advertisements
Similar presentations
Lecture III: Collective Behavior of Multi -Agent Systems: Analysis
Advertisements

Lecture V: Game Theory Zhixin Liu Complex Systems Research Center,
DYNAMICS OF RANDOM BOOLEAN NETWORKS James F. Lynch Clarkson University.
Week 11 Review: Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution.
Evolution and Repeated Games D. Fudenberg (Harvard) E. Maskin (IAS, Princeton)
Evolution of Cooperation The importance of being suspicious.
An Introduction to... Evolutionary Game Theory
VL Netzwerke, WS 2007/08 Edda Klipp 1 Max Planck Institute Molecular Genetics Humboldt University Berlin Theoretical Biophysics Networks in Metabolism.
Maynard Smith Revisited: Spatial Mobility and Limited Resources Shaping Population Dynamics and Evolutionary Stable Strategies Pedro Ribeiro de Andrade.
Prisoner’s dilemma TEMPTATION>REWARD>PUNISHMENT>SUCKER.
Evolving New Strategies The Evolution of Strategies in the Iterated Prisoner’s Dilemma 01 / 25.
Zhixin Liu Complex Systems Research Center,
280 SYSTEM IDENTIFICATION The System Identification Problem is to estimate a model of a system based on input-output data. Basic Configuration continuous.
1 Lecture 8: Genetic Algorithms Contents : Miming nature The steps of the algorithm –Coosing parents –Reproduction –Mutation Deeper in GA –Stochastic Universal.
Data Mining Techniques Outline
1. Elements of the Genetic Algorithm  Genome: A finite dynamical system model as a set of d polynomials over  2 (finite field of 2 elements)  Fitness.
Evolutionary Games The solution concepts that we have discussed in some detail include strategically dominant solutions equilibrium solutions Pareto optimal.
Genetic Algorithms and Their Applications John Paxton Montana State University August 14, 2003.
Advanced Topics in Data Mining Special focus: Social Networks.
COMP305. Part II. Genetic Algorithms. Genetic Algorithms.
6/4/03Genetic Algorithm The Genetic Algorithm The Research of Robert Axelrod The Algorithm of John Holland Reviewed by Eyal Allweil and Ami Blonder.
Intro to AI Genetic Algorithm Ruth Bergman Fall 2004.
CHAPTER 12 ADVANCED INTELLIGENT SYSTEMS © 2005 Prentice Hall, Decision Support Systems and Intelligent Systems, 7th Edition, Turban, Aronson, and Liang.
MAKING COMPLEX DEClSlONS
Genetic Algorithm.
Genetic Algorithms and Ant Colony Optimisation
The Multiplicative Weights Update Method Based on Arora, Hazan & Kale (2005) Mashor Housh Oded Cats Advanced simulation methods Prof. Rubinstein.
Learning in Multiagent systems
Slides are based on Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems.
Soft Computing Lecture 18 Foundations of genetic algorithms (GA). Using of GA.
1 Local search and optimization Local search= use single current state and move to neighboring states. Advantages: –Use very little memory –Find often.
CS 484 – Artificial Intelligence1 Announcements Lab 3 due Tuesday, November 6 Homework 6 due Tuesday, November 6 Lab 4 due Thursday, November 8 Current.
Zorica Stanimirović Faculty of Mathematics, University of Belgrade
Example Department of Computer Science University of Bologna Italy ( Decentralised, Evolving, Large-scale Information Systems (DELIS)
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
Presenter: Chih-Yuan Chou GA-BASED ALGORITHMS FOR FINDING EQUILIBRIUM 1.
ART – Artificial Reasoning Toolkit Evolving a complex system Marco Lamieri
1/27 Discrete and Genetic Algorithms in Bioinformatics 許聞廉 中央研究院資訊所.
Part 1: Biological Networks 1.Protein-protein interaction networks 2.Regulatory networks 3.Expression networks 4.Metabolic networks 5.… more biological.
Evolving Virtual Creatures & Evolving 3D Morphology and Behavior by Competition Papers by Karl Sims Presented by Sarah Waziruddin.
Daniel Ariosa Ecole Polytechnique Fédérale de Lausanne (EPFL) Institut de Physique de la Matière Complexe CH-1015 Lausanne, Switzerland and Hugo Fort Instituto.
10 December, 2008 CIMCA2008 (Vienna) 1 Statistical Inferences by Gaussian Markov Random Fields on Complex Networks Kazuyuki Tanaka, Takafumi Usui, Muneki.
2005MEE Software Engineering Lecture 11 – Optimisation Techniques.
Algorithms and their Applications CS2004 ( ) 13.1 Further Evolutionary Computation.
Evolution Programs (insert catchy subtitle here).
Edge Assembly Crossover
Section 2 – Ec1818 Jeremy Barofsky
1. Genetic Algorithms: An Overview  Objectives - Studying basic principle of GA - Understanding applications in prisoner’s dilemma & sorting network.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Game Theory by James Crissey Luis Mendez James Reid.
Iterated Prisoner’s Dilemma Game in Evolutionary Computation Seung-Ryong Yang.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
D Nagesh Kumar, IIScOptimization Methods: M8L5 1 Advanced Topics in Optimization Evolutionary Algorithms for Optimization and Search.
Evolving Strategies for the Prisoner’s Dilemma Jennifer Golbeck University of Maryland, College Park Department of Computer Science July 23, 2002.
Genetic Algorithm Dr. Md. Al-amin Bhuiyan Professor, Dept. of CSE Jahangirnagar University.
Artificial Intelligence By Mr. Ejaz CIIT Sahiwal Evolutionary Computation.
Computacion Inteligente Least-Square Methods for System Identification.
Evolution of Cooperation in Mobile Ad Hoc Networks Jeff Hudack (working with some Italian guy)
Genetic Algorithms And other approaches for similar applications Optimization Techniques.
March 1, 2016Introduction to Artificial Intelligence Lecture 11: Machine Evolution 1 Let’s look at… Machine Evolution.
Presented By: Farid, Alidoust Vahid, Akbari 18 th May IAUT University – Faculty.
 Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems n Introduction.
Cmpe 588- Modeling of Internet Emergence of Scale-Free Network with Chaotic Units Pulin Gong, Cees van Leeuwen by Oya Ünlü Instructor: Haluk Bingöl.
Game Theory and Cooperation
Evolving New Strategies
Artificial Intelligence (CS 370D)
Subject Name: Operation Research Subject Code: 10CS661 Prepared By:Mrs
Metaheuristic methods and their applications. Optimization Problems Strategies for Solving NP-hard Optimization Problems What is a Metaheuristic Method?
Introduction to Artificial Intelligence Lecture 11: Machine Evolution
Boltzmann Machine (BM) (§6.4)
Presentation transcript:

Lecture VI: Adaptive Systems Zhixin Liu Complex Systems Research Center, Academy of Mathematics and Systems Sciences, CAS

In the last lecture, we talked about Game Theory An embodiment of the complex interactions among individuals Nash equilibrium Evolutionarily stable strategy

In this lecture, we will talk about Adaptive Systems

Adaptation To adapt: to change oneself to conform to a new or changed circumstance. What we know from the new circumstance? Adaptive estimation, learning, identification How to do the corresponding response? Control/decision making

Why Adaptation? Uncertainties always exist in modeling of practical systems. Adaptation can reduce the uncertainties by using the system information. Adaptation is an important embodiment of human intelligence.

Framework of Adaptive Systems Environment system control

Two levels of adaptation Individual: learn and adapt Population level Death of old individuals Creation of new individuals Hierarchy

Some Examples Adaptive control systems Iterated prisoner’s dilemma adaptation in a single agent Iterated prisoner’s dilemma adaptation among agents

Some Examples Adaptive control systems Iterated prisoner’s dilemma adaptation in a single agent Iterated prisoner’s dilemma adaptation among agents

Adaptation In A Single Agent Environment system control wt yt ut

Information= prior+ posterior wt ut yt Dynamical System Information= prior+ posterior =I0+I1 I0 = prior knowledge about the system I1 = posterior knowledge about the system ={u0,u1,…ut, y0,y1,…,yt} (Observations) I The posterior information can be used to reduce the uncertainties of the system.

Uncertainty Model External uncertainty: noise/disturbance Internal Uncertainty External uncertainty: noise/disturbance Internal uncertainty: Parameter uncertainty Signal uncertainty Functional uncertainty

Adaptation To adapt: to change oneself to conform to a new or changed circumstance. What we know from the new circumstance? Adaptive estimation, learning, identification How to do the corresponding response? Control/decision making

Adaptive Estimation

Adaptive Estimation Adaptive estimation: parameter or structure estimator, which can be updated based on the on-line observations. ŷt Adaptive Estimator e - + ∑ yt ut System Example: In the parametric case, the parameter estimator can be obtained by minimizing certain prediction error:

Adaptive Estimation Parameter estimation : unknown parameter vector Consider the following linear regression model: : unknown parameter vector : regression vector : noise sequence Remark Linear regression model may be nonlinear. Linear system can be translated into linear regression model.

Least Square (LS) Algorithm 1795, Gauss, least square algorithm The number of functions is greater than that of the unknown parameters. The data contain noise. Minimize the following prediction error:

Recursive Form of LS Recursive Form of LS: where Pt is the following estimation “covariance” matrix A basic problem:

Recursive Form of LS Assumption 1: 1) The noise sequence is a martingale difference sequence, and there exists a constant , such that 2) The regression vector is an adaptive sequence, i.e., Theorem (T.L. Lai & C.Z. Wei) Under the above assumption, if the following condition holds then the LS has the strong consistency.

Weighted Least Square Minimize the following prediction error: Recursive form of WLS:

Self-Convergence of WLS Take the weight as follows: with . Theorem Under Assumption 1, for any initial value and any regression vector , will converge to some vector almost surely. Lei Guo, 1996, IEEE TAC

Adaptation To adapt: to change oneself to conform to a new or changed circumstance. What we know from the new circumstance? Adaptive estimation, learning, identification How to do the corresponding response? Control/decision making

Adaptive Control

Adaptive Control Adaptive Control: a controller with adjustable parameters (or structures) together with a mechanism for adjusting them. y u Adaptive Estimator Plant r Adaptive Controller r

Robust Control Model = Nominal +”Ball” r Can not reduce uncertainty!

Adaptive Control An example Consider the following linear regression model: Where a and b are unknown parameters, yt , ut, and wt are the output, input and white noise sequence. Objective: design a control law to minimize the following average tracking errors

Adaptive Control If (a,b) is known, we can get the optimal controller: “Certainty Equivalence” Principle: Replace the unknown parameters in a non-adaptive controller by its online estimate If (a,b) is unknown, the adaptive controller can be taken as

Adaptive control If (a,b) is unknown, the adaptive controller can be taken as with where (at,bt) can be obtained by LS:

Adaptive Control The closed-loop system:

Theoretical Problems a) Stability: b) Optimality:

Theoretical Obstacles Controller Closed-loop system Estimation Data

Theoretical Obstacles 1) The closed-loop system is a very complicated nonlinear stochastic dynamical system. 2) No useful statistical properties, like stationarity or independency of the system signals are available. 3) No properties of (at, bt) are known a priori.

Theorem Assumption: 1) The noise sequence is a martingale difference sequence, and there exists a constant , such that 2) The regression vector is an adaptive sequence, i.e., 3) is a deterministic bounded signal. Theorem Under the above assumptions, the closed-loop system is stable and optimal. Lei Guo, Automatica, 1995

Some Examples Adaptive control systems Iterated prisoner’s dilemma adaptation in a single agent Iterated prisoner’s dilemma adaptation among agents

Prisoner’s Dilemma (3,3) (0,5) (1,1) (5,0) The story of prisoner’s dilemma Player: two prisoners Action: {cooperation, Defect} Payoff matrix Prisoner B C D (3,3) (0,5) C Prisoner A (5,0) (1,1) D

Prisoner’s Dilemma (3,3) (0,5) (1,1) (5,0) No matter what the other does, the best choice is “D”. (D,D) is a Nash Equilibrium. But, if both choose “D”, both will do worse than if both select “C” Prisoner B C D (3,3) (0,5) C Prisoner A (5,0) (1,1) D

Iterated Prisoner’s Dilemma The individuals: Meet many times Can recognize a previous interactant Remember the prior outcome Strategy: specify the probability of cooperation and defect based on the history P(C)=f1(History) P(D)=f2(History)

Strategies Tit For Tat – cooperating on the first time, then repeat opponent's last choice. Player A C D D C C C C C D D D D C… Player B D D C C C C C D D D D C… Tit For Tat and Random - Repeat opponent's last choice skewed by random setting.* Tit For Two Tats and Random - Like Tit For Tat except that opponent must make the same choice twice in a row before it is reciprocated. Choice is skewed by random setting.* Tit For Two Tats - Like Tit For Tat except that opponent must make the same choice twice in row before it is reciprocated. Naive Prober (Tit For Tat with Random Defection) - Repeat opponent's last choice (ie Tit For Tat), but sometimes probe by defecting in lieu of cooperating.* Remorseful Prober (Tit For Tat with Random Defection) - Repeat opponent's last choice (ie Tit For Tat), but sometimes probe by defecting in lieu of cooperating. If the opponent defects in response to probing, show remorse by cooperating once.* Naive Peace Maker (Tit For Tat with Random Co-operation) - Repeat opponent's last choice (ie Tit For Tat), but sometimes make peace by co-operating in lieu of defecting.* True Peace Maker (hybrid of Tit For Tat and Tit For Two Tats with Random Cooperation) - Cooperate unless opponent defects twice in a row, then defect once, but sometimes make peace by cooperating in lieu of defecting.* Random - always set at 50% probability

Strategies Tit For Tat – cooperating on the first time, then repeat opponent's last choice. Player A C D D C C C C C D D D D C… Player B D D C C C C C D D D D C… Tit For Tat and Random - Repeat opponent's last choice skewed by random setting.* Tit For Two Tats and Random - Like Tit For Tat except that opponent must make the same choice twice in a row before it is reciprocated. Choice is skewed by random setting.* Tit For Two Tats - Like Tit For Tat except that opponent must make the same choice twice in row before it is reciprocated. Naive Prober (Tit For Tat with Random Defection) - Repeat opponent's last choice (ie Tit For Tat), but sometimes probe by defecting in lieu of cooperating.* Remorseful Prober (Tit For Tat with Random Defection) - Repeat opponent's last choice (ie Tit For Tat), but sometimes probe by defecting in lieu of cooperating. If the opponent defects in response to probing, show remorse by cooperating once.* Naive Peace Maker (Tit For Tat with Random Co-operation) - Repeat opponent's last choice (ie Tit For Tat), but sometimes make peace by co-operating in lieu of defecting.* True Peace Maker (hybrid of Tit For Tat and Tit For Two Tats with Random Cooperation) - Cooperate unless opponent defects twice in a row, then defect once, but sometimes make peace by cooperating in lieu of defecting.* Random - always set at 50% probability

Strategies Always Defect Always Cooperate Grudger (Co-operate, but only be a sucker once) - Cooperate until the opponent defects. Then always defect unforgivingly. Pavlov (repeat last choice if good outcome) - If 5 or 3 points scored in the last round then repeat last choice. Pavlov / Random (repeat last choice if good outcome and Random) - If 5 or 3 points scored in the last round then repeat last choice, but sometimes make random choices.* Adaptive - Starts with c,c,c,c,c,c,d,d,d,d,d and then takes choices which have given the best average score re-calculated after every move. Gradual - Cooperates until the opponent defects, in such case defects the total number of times the opponent has defected during the game. Followed up by two co-operations. Suspicious Tit For Tat - As for Tit For Tat except begins by defecting. Soft Grudger - Cooperates until the opponent defects, in such case opponent is punished with d,d,d,d,c,c. Customised strategy 1 - default setting is T=1, P=1, R=1, S=0, B=1, always co-operate unless sucker (ie 0 points scored). Customised strategy 2 - default setting is T=1, P=1, R=0, S=0, B=0, always play alternating defect/cooperate.

Iterated Prisoner’s Dilemma Which strategy can thrive/what is the good strategy? Robert Axelrod, 1980s A computer round-robin tournament The first round The second round AXELROD R. 1987. The evolution of strategies in the iterated Prisoners' Dilemma. In L. Davis, editor, Genetic Algorithms and Simulated Annealing. Morgan Kaufmann, Los Altos, CA.

Characters of “good” strategies Goodness: never defect first First round: the first eight strategies with “goodness” Second round: fourteen strategies with “goodness” in the first fifteen strategies Forgiveness: may revenge, but the memory is short. “Grudger” is not s strategy with “forgiveness” “Goodness” and “forgiveness” is a kind of collective behavior. For a single agent, defect is the best strategy.

Evolution of the Strategies Evolve “good” strategies by genetic algorithm (GA)

Some Notations in GA String: the individuals, and it is used to represent the chromosome in genetics. Population: the set of the individuals Population size: the number of the individuals Gene: the elements of the string E.g., S=1011, where 1,0,1,1 are called genes. Fitness: the adaptation of the agent for the circumstance From Jing Han’s PPT

How GA works? Represent the solution of the problem by “chromosome”, i.e., the string Generate some chromosomes as the initial solution randomly According to the principle of “Survival of the Fittest ”, the chromosome with high fitness can reproduce, then by crossover and mutation the new generation can be generated. The chromosome with the highest fitness may be the solution of the problem. From Jing Han’s PPT

GA choose an initial population Natural Selection Create new generation crossover choose an initial population determine the fitness of each individual perform selection repeat perform crossover perform mutation until some stopping criterion applies mutation From Jing Han’s PPT

Some Remarks On GA The GA search the optimal solution from a set of solution, rather than a single solution The search space is large: {0,1}L GA is a random algorithm, since selection, crossover and mutation are all random operations. Suitable for the following situation: There is structure in the search space but it is not well-understood The inputs are non-stationary (i.e., the environment is changing) The goal is not global optimization, but finding a reasonably good solution quickly

Evolution of Strategies By GA Each chromosome represents one strategy The strategy is deterministic and it is determined by the previous moves. E.g., the strategy is determined by one step history, then there are four cases of history Player I C D D C Player II D D C C The number of the possible strategies is 2*2*2*2=16. TFT: F(CC)=C, F(CD)=D, F(DC)=C, F(DD)=D Always cooperate: F(CC)=F(CD)=F(DC)=F(DD)=C Always defect: F(CC)=F(CD)=F(DC)=F(DD)=D …

Evolution of the Strategies Strategies: use the outcome of the three previous moves to determine the current move. The possible number of the histories is 4*4*4=64. Player I CCC CCD CDC CDD DCC DCD … DDD DDD Player II CCC CCC CCC CCC CCC CCC … DDC DDD C C C C C C … C C C C C C C C … C D D D D D D D … D D The initial premises is three hypothetical move. The length of the chromosome is 70. The total number of strategies is 270≈1021.

Evolution of “good” strategy Five steps of evolving “good” strategies by GA An initial population is chosen. Each individual is run in the current environment to determine its effectiveness. The relatively successful individual are selected to have more offspring. The successful individuals are randomly paired off to produce two offspring per mating. Crossover: way of constructing the chromosomes of the two offspring from the chromosome of two parents. Mutation: randomly changing a very small proportion of the C’s to D’s and vice versa. New population are generated.

Evolution of the Strategies Some parameters: The population size in each generation is 20. Each game consists of 151 moves. Each of them meet eight representatives, and this made about 24,000 moves per generation. A run consists of 50 generation Forty runs were conducted.

Results The median member is as successful as TFT Most of the strategies is resemble TFT Some of them have the similar patterns as TFT Do not rock the boat: continue to cooperate after the mutual cooperation Be provocable: defect when the other player defects out of the blue Accept an apology: continue to cooperate after cooperation has been restored Forget: cooperate when mutual cooperation has been restored after an exploitation Accept a rut: defect after three mutual defections

What is a “good” strategy? TFT is a good strategy? Tit For Two Tats may be the best strategy in the first round, but it is not a good strategy in the second round. “Good” strategy depends on other strategies, i.e., environment. Evolutionarily stable strategy

Evolutionarily stable strategy (ESS) Introduced by John Maynard Smith and George R. Price in 1973 ESS means evolutionarily stable strategy, that is “a strategy such that, if all member of the population adopt it, then no mutant strategy could invade the population under the influence of natural selection.” ESS is robust for evolution, it can not be invaded by mutation. John Maynard Smith, “Evolution and the Theory of Games”

Definition of ESS A strategy x is an ESS if for all y, y  x, such that holds for small positiveε.

ESS in IPD Tit For Tat can not be invaded by the wiliness strategies, such as always defect. TFT can be invaded by “goodness” strategies, such as “always cooperate”, “Tit For Two Tats” and “Suspicious Tit For Tat ” Tit For Tat is not a strict ESS. “Always Cooperate” can be invaded by “Always Defect”. “Always Defect ” is an ESS.

Other Adaptive Systems Complex adaptive system John Holland, Hidden Order, 1996 Examples: stock market, social insect, ant colonies, biosphere, brain, immune system, cell , developing embryo, … Evolutionary algorithms genetic algorithm, neural network, …

References Lei Guo, Self-convergence of weighted least-squares with applications to stochastic adaptive control, IEEE Trans. Automat. Contr., 1996, 41(1): 79-89. Lei Guo, Convergence and logarithm laws of self-tuning regulators, 1995, Automatica, 31(3): 435-450. Lei Guo, Adaptive systems theory: some basic concepts, methods and results, Journal of Systems Science and Complexity, 16(3): 293-306. Drew Fudenberg, Jean Tirole, Game Theory, The MIT Press, 1991. AXELROD R. 1987, The evolution of strategies in the iterated Prisoners' Dilemma. In L. Davis, editor, Genetic Algorithms and Simulated Annealing. Morgan Kaufmann, Los Altos, CA. Richard Dawkins, The Selfish Gene, Oxford University Press. John Holland, Hidden Order, 1996.

Adaptation in a single agent Adaptation in games Adaptation in a single agent

Summary In these six lectures, we have talked about: Complex Networks Collective Behavior of MAS Game Theory Adaptive Systems

Summary In these six lectures, we have talked about: Complex Networks: Topology Collective Behavior of MAS Game Theory Adaptive Systems

Three concepts Short average path length Large clustering coefficient where dij is the shortest distance between i and j. Clustering Coefficient C=<C(i)> Degree distribution P(k)=probability that the randomly chosen node i has exactly k neighbors Short average path length Large clustering coefficient Power law degree distribution

Regular Graphs Regular graphs: graphs where each vertex has the same number of neighbors. Examples: complete graph ring graph lattice

Random Graph ER random graph model G(N,p) Given N nodes Add an edge between a randomly-selected pair of nodes with probability p Homogeneous nature: each node has roughly the same number of edges

Small World Networks WS model Introduce pNK/2 long-range edges A few long-range links are sufficient to decrease l, but will not significantly change C.

Scale Free Networks Some observations A breakthrough: Barabási & Albert, 1999, Science Generating process of BA model: 1) Starting with a network with m0 nodes 2) Growth: at each step, we add a new node with m(≦m0) edges that link the new node to m different nodes already present in the network. 3) Preferential attachment: When choosing the nodes to which the new nodes connects, we assume that the probability ∏ that a new node will be connected to node i on the degree ki of node i, such that

Summary In these six lectures, we have talked about: Complex Networks: Topology Collective Behavior of MAS: More is different Game Theory Adaptive Systems

Multi-Agent System (MAS) Many agents Local interactions between agents Collective behavior in the population level More is different.---Philp Anderson, 1972 e.g., Phase transition, coordination, synchronization, consensus, clustering, aggregation, …… Examples: Physical systems Biological systems Social and economic systems Engineering systems … …

Vicsek Model Neighbors: Position: Heading:

Synchronization of the linearized Vicsek model Theorem 2 (Jadbabaie et al. , 2003) Joint connectivity of the neighbor graphs on each time interval [th, (t+1)h] with h >0 Synchronization of the linearized Vicsek model Related result: J.N.Tsitsiklis, et al., IEEE TAC, 1984

Theorem 7 High Density Implies Synchronization For any given system parameters and when the number of agnets n is large, the Vicsek model will synchronize almost surely. This theorem is consistent with the simulation result.

Theorem 8 High density with short distance interaction Let and the velocity satisfy Then for large population, the MAS will synchronize almost surely.

Soft Control Key points: U(t) y(t) Different from distributed control approach. Intervention to the distributed system Not to change the local rule of the existing agents Add one (or a few) special agent – called “shill” based on the system state information, to intervene the collective behavior; The “ shill” is controlled by us, but is treated as an ordinary agent by all other agents. Shill is not leader, not leader-follower type. Feedback intervention by shill(s). This page is very important! From Jing Han’s PPT

Leader-Follower Model Ordinary agents Information agents Key points: Not to change the local rule of the existing agents. Add some (usually not very few) “information” agents – called “leaders”, to control or intervene the MAS; But the existing agents treated them as ordinary agents. The proportion of the leaders is controlled by us (If the number of leaders is small, then connectivity may not be guaranteed). Open-loop intervention by leaders.

Summary In these six lectures, we have talked about: Complex Networks: Topology Collective Behavior of MAS: More is different Game Theory: Interactions Adaptive Systems

Definition of Nash Equilibrium Nash Equilibrium (NE): A solution concept of a game (N, S, u) : a game Si: strategy set for player i : set of strategy profiles : payoff function s-i: strategy profile of all players except player i A strategy profile s* is called a Nash equilibrium if where σi is any pure strategy of the player i. For a game composed of N players, S is the strategy set formed by all agents, u: is the payoff function, S_i is the strategy set fotr player i. S Since the payoff matrix is open to all players, so for each player can predict the strategy that his opponents will adopt, All players can forecast a specified equilibrium, then there is no motivation to adopt a different strategy for each player.

Definition of ESS A strategy x is an ESS if for all y, y  x, such that holds for small positiveε.

Summary In these six lectures, we have talked about: Complex Networks: Topology Collective Behavior of MAS: More is different Game Theory: Interactions Adaptive Systems: Adaptation

Other Topics … Self-organizing criticality Nonlinear dynamics Earthquakes, fire, sand pile model, Bak-Sneppen model … Nonlinear dynamics chaos, bifurcation, Artificial life Tierra model, gene pool, game of life,… Evolutionary dynamics genetic algorithm, neural network, … …

Complex systems Not a mature subject No unified framework or universal methods

THE END