6 Evolutionary Game Theory & Replicator Dynamics Flashcards
CGT vs EGT
While Classical Game Theory (CGT) is a static theory centred around players’ full rationality, CHOOSING MOST RATIONAL CHOICE Evolutionary Game Theory (EGT) is about dynamic adaptation in games played between
randomly paired individuals, without consideration of rationality and reciprocity (no information about previous games).
EGT and population games
Evolutionary dynamics can be applied to many systems. Here, we focus on
describes the evolution of interacting and generally competing populations
symmetric
two-player evolutionary games and adopt the framework of population games
interactions between different species= each pure strategy gives payoff matrix A
frequencies = densities of individuals
how they fare depends on their fitness = expected payoff
fitness
defined as the reproductive potential of an individual.
population games
each pure strategy i=1,…,k is interpreted as one of k species composing the population
frequency x_i of pure strategy i is the fraction x_i of species i in the population
we assume pop is very large and well mixed (no fluctuations and no space)
spatially homogeneous
sum x_i = 1 from i=1,…,k as is fraction
strategy profile
at each time t, the state of the population (COMPOSITION) is described by “strategy profile”
x(t) = (x₁(t),x₂(t),…,xₖ(t))^T
=x₁(t)e₁+x₂(t)e₂+…xₖ(t)eₖ
=
(x₁(t))
(x₂(t))
(….)
(xₖ(t))
e₁=(1,0,…,0)^T
e₂= (0,1,…,0)^T
…
eₖ=(0,0,..,1)^T
represent the k pure strategies / species
individuals
each individual is associated with a pure strategy, and at each time increment, agents are randomly paired and play a stage game their expected payoffs determine the rate of replication.
k species
k pure strategies
expected payoff
Π(eᵢ,x)
≡ Πᵢ(x)
= eᵢ^T Ax
with k-dimensional unit vectors
e₁=(1,0,…,0)^T
e₂=(0,1,0,…,0)^T etc
Πᵢ(x) expected payoff of an i-player against a random individual of the population playing on average x
replicator
The basic ingredient in an evolutionary system
an entity (gene, organism, strategy,,,,,)
able to produce copies of itself. A replicator system is a set of replicators with a pattern of interactions among individuals, such that those with higher fitness reproduce faster
WHEN xᵢ(t)=0,1
population only one species
MONOMORPHIC
1 pure strategy in CGT
WHEN 0< xᵢ(t)<1
population consists of the mixing of more than 1 species, is said “polymorphic”
mixed strategy in CGT)
genome and strategies
n a biological setting, we assume that the strategies are encoded by the genome, and
there is no assumption about the rationality of players. In close relation to population
genetics, the success of one species in EGT is measured by its fitness which depends on
what others do: it changes with the population composition as prescribed by the stage
game. The evolutionary dynamics of x(t) in a large population is usually specified by
the replicator equations, see Sec. 6.2
CGT and EGT interpretation of pop games
CGT EGT
Utility function, expected payoff VS Fitness, reproductive potential
Perfect rationality assumed VS Rationality not assumed
Static theory VS Dynamic theory
Pure strategy i = 1, . . . , k VS Species i = 1, . . . , k
Pure strategy i ⇒ unit vector ei VS Species i ⇒ unit vector e_i
Strategy profile x =sumi=1 to k x_ie_i VS Population state: x(t) = (x1(t), . . . , xk(t))T
Frequency of pure strategy i: 0 ≤ xi ≤ 1 xi(t): VS fraction of i in the population at t
sum_i=1,k x_i = 1 x_i(t) VS changes in time but sum_i=1,k x_i(t) = 1
EXAMPLE
Here, there are 3 species and 10 individuals,
4 of species 1, 3 of species 2 and 3 of species 3
x_1=4/10
x_2=3/10
x_3=3/10
representing with unit vectors respectively
(1,0,0)^T
(0,1,0)^T
(0,0,1)^T
x=0.4e₁+0.3e₂+0.3e₃ =
(0.4,0.3,0.3)^T
How does the population, i.e. , evolve?
we use replicator dynamics to see
REPLICATOR DYNAMICS
EGT doesn’t rely on rationality but considers a large population of individuals with given strategies that interact randomly in games
assume that between a time increment
(between t and t +dt) population composition is x(t)
individuals meet randomly many times in pairwise contests for payoffs- finesses and encode reproductive potential
replicator dynamics:
Between time t and t+dt, the population is in state
x(t)=
(x_1(t)
Population’s
average fitness
At this point, individuals are paired randomly to play “stage games” many times (each time
starting from scratch, without recollection of previous outcomes) => they obtain their expected payoff = “fitness” Π_i(x(t)))
now it depends on x at time t
and the population average payoff/fitness
Π_bar(x(t))
Fitness For species i:
Πᵢ(x(t))= eᵢ^T A x(t)
says how i is expected to perform when the population is state x(t)
Population’s
average fitness:
Π_barᵢ(x(t))= x^T A x(t)
says how a random individual perform on average when the
population is state x(t)
(these are, generally change in time: they are “frequency dependent”)
Darwinian principle
the best adapted species/strategies reproduce faster than others =>
compare the fitness of each pure strategy
we compare the fitness of each species i to the population average fitness
Π_barᵢ(x(t)) giving growth rate
Growth rate
at which its density xᵢ(t) changes in time,
is given by
x_dot ᵢ ≡ dxᵢ(t)/dt
x_dot ᵢ/xᵢ = (d/dt)(ln xᵢ(t))
where the GR is generally a function of t, such that
x_i(t) ~ exp(∫ᵗ GR(t′)dt′) with x_i(t) that increases when GR >0 and decreases when GR<0
pop grows exponentially for a short time
GR=fitness of species i - population average fitness
=Πᵢ- Π_barᵢ
species i’s growth rate
Π_ᵢ(x(t)) - Π_barᵢ(x(t))
pop grows exponentially for a short time
GR=fitness of species i - population average fitness
=Πᵢ- Π_barᵢ
species i’s growth rate
xᵢ increases when Πᵢ(x(t))> Π_barᵢ(x(t))
species i has a higher fitness than average population
growth
xᵢ decreases when Πᵢ(x(t))< Π_barᵢ(x(t))
if species i has a lower fitness than average population
decrease
THIS LEADS TO THE REPLICATOR DYNAMICS
REPLICATOR EQUATIONS
THIS LEADS TO THE REPLICATOR DYNAMICS A
replicator equations
xᵢ_dot = xᵢ [ Πᵢ(x)-Π_barᵢ(x)] =xᵢ[ i’s growth rate]
= xᵢ[ eᵢ^T Ax- x^T Ax] (unit vector = Π_bar)
=xᵢ[(Ax)ᵢ- x^TAx] for i=1,…,k
set of k coupled nonlinear ODEs
since sum x_i=1
=> k-1 independent coupled nonlinear ODEs, e.g. by choosing to write x_k=1-(…)
case k=2 is nice: scalar ode
Use the second format
At each time increment dt, the population state changes according to the REs:
xᵢ →xᵢ+dxᵢ
where
dxᵢ=xᵢ[Πᵢ(x)-Π_bar(x)] dt
Properties of the REs (symmetric games):
- REs are a set of k-1 coupled nonlinear equations ODES for (generally cubic in x_i, )
- A symmetric zero-sum game can be formulated so that one player gains what the other loses
=> can be chosen to be antisymmetric and, in this formulation, for zero-sum games
=> For these zero-sum games, REs are quadratic: x_dot ᵢ = xᵢ(**Ax)ᵢ
=> When these REs have a physical interior equilibrium, there is a nontrivial constant of motion (like in the zero-sum RPS game of Sections 5.6 and 6.3, constant of motions, orbits set by IC)
Each unit vector eᵢ=(0,…,0,i,0…,0^T (0 x (i-1), 1, (0 x (k-i)) associated with pure strategy i, is an equilibrium of the REs corresponding to the entire population consisting of individuals of species i.
These equilibria are called “absorbing fixed points” because there is no longer species coexistence
(loss of variability- homogeneous pop) in these states. They are particularly important in finite populations. (See Chapter 9)
The REs have at most 1 interior equilibrium
x∗ = (x∗1, x∗2, . . . , x∗k)^T
with 0<x∗_i<1, for which a necessary condition is
(Ax∗)₁= (Ax∗)₂=….
= (Ax∗)ₖ = x∗^TAx∗ (same conditions as for BCT, demanding that what is in the square brackets vanishes, each component is the same coinciding with avg. fitness)
Adding (or subtracting) a constant to each column of A gives the same REs, i.e
A=
[a b]
[c d] and
A~ =
[ a-c 0]
[0 d-b]
lead to the same REs
Replicator dynamics with or result in the same replicator equations.
We can obtain same RE by adding subtracting
replicator dynamics summary
in terms of ODES dep on evolution of classical game theory,
interpret expected payoff as fitness
growth rate relates to difference in fitness
Classical game theory (CGT): static theory assuming full rationality of all players.
Evolutionary game theory (EGT): theory of dynamic adaptation assuming no rationality, replaced by the notion of fitness (expected payoff) => framework for evolution of interacting populations
Replicator eqs
for frequency xᵢ of pure strategy i for symmetric 2-player game
payoff matrix
time derivative for fraction x_i
x˙ᵢ = xᵢ[Πᵢ(x) − Π¯(x)]
= xᵢ[(Ax)ᵢ − xT Ax] (for i = 1, . . . , k)
xₖ = 1 − Σᵢ₌₁ ᵏ⁻¹ xᵢ
=>k-1 indep. variables
set of coupled non linear ODEs
k pure strategies become k-1 indep vars
k=2 gives one ODE
What is the relationship between the notion of stability in nonlinear dynamics
and the concepts of evolutionary stability and Nash equilibrium?
notions are close but do not perfectly overlap, which gives rise to what is referred to as
“Folk theorems”.
given ODEs we can find equilibria,
we saw that all equilibria correspond to a pure strategy of each species taking over
we also saw that there can be 0 or 1 coexistence equilirbium, by demanding [*] to vanish, relates to BCT
x˙ᵢ = xᵢ[Πᵢ(x) − Π¯(x)]
= xᵢ[(Ax)ᵢ − xT Ax]
= xᵢ[*]
Summary relations between
dynamic and evolutionary stability:
(a) NE of the underlying stage game= equilibria of the REs , while strict NE of the underlying stage game are attractors (asymptotically equilibria) of the REs.
(b) A stable equilibrium of the REs is a NE of the underlying stage game.
(c) If a solution of the REs converges to some x, then x is a NE of the stage game.
(d) The evolutionary stable strategy ESSes of the stage game is an attractor of the REs
Moreover, interior ESSes (see Theorem 5.2) of the underlying stage games are global
attractors of the REs (6.4), i.e. their basin of attraction is the whole simplex S.
Importantly, the converse of statements (a)-(d) are not true. Proofs, along with examples
and counter-examples, can, e.g., be found in Ref.(4) (see Sec. 6.6).
(e) For two-player symmetric matrix games with two pure strategies (2×2 symmetric
games), things are fortunately very simple: in this case, x is an ESS if and only if it
is an attractor of the REs (6.4).
Summary relations between
dynamic and evolutionary stability:
- A Nash equilibrium (NE) of a stage game is an equilibrium of the associated replicator equations (REs)
- Strict Nash equilibria are attractors(asympotic stable equilibria) of the associated REs
- A stable equilibrium of the REs is NE of the underlying stage game
- Evolutionary stable strategies (ESSes) of a stage game are attractors of the underlying REs
Note: converse statements are not true! (e.g. attractors of the REs are not ESSes of the stage game)
Relation is much simpler for symmetric 2-player games with 2 pure strategies:
In this special, but important case:
ESSes <=> ATTRACTORS OF THE REs
Replicator dynamics of two-player symmetric games with two pure strategies
symmetric two-player games with two pure
strategies (2 × 2 games)
REs are scalar ODEs (:D)
infinitely large and
unstructured population where individuals (players) have two pure strategies at their disposal, and are randomly paired to play the underlying symmetric 2 × 2 stage game.
replicator dynamics:
symmetric two-player games with two pure
strategies (2 × 2 games)
payoff matrix
frequency
“cooperation” (C)
e𝒸 = (1, 0)^T
“defection” (D)
e𝒹 = (0, 1)^T
A=
vs |C |D
C a b
D c d
Frequency of species/strategy C is x, frequency of species/strategy D is 1-x
(Fractions)
replicator dynamics:
symmetric two-player games with two pure strategies (2 × 2 games)
Strategy profile
i.e. state of the population, at time t:
x(t) = (x(t), 1 − x(t))^T
= x(t)e𝒸 + (1 − x(t))e𝒹
replicator dynamics:
symmetric two-player games with two pure strategies (2 × 2 games)
expected payoffs
Average payoff/fitness
= xΠ𝒸 + (1 − x)Π𝒹
Expected payoff of C = fitness of C in state
x(t) : Π𝒸 = e𝒸^TAx(t)
= ax(t) + b(1 − x(t))
=(1 0) A ( x, 1-x)^T
playing c against x
(note x is x(t))
Expected payoff of D = fitness of D in state x(t) : Π𝒹 = e𝒹^TAx(t)
= cx(t) + d(1 − x(t))
=(1 0) A ( x, 1-x)^T
avrg payoff/fitness:
Π¯(t) = x(t)^T Ax(t)
= x(t)Π𝒸(t) + (1 − x(t))Π𝒹(t)
=(x, 1-x) A (x, 1-x)
=(x 1 − x) (Π𝒸)
(Π𝒹)
[ax + b(1 − x)]
[cx + d(1 − x)]
=
(Π𝒸)
(Π𝒹)
x is the fraction of cooperation in pop x Π𝒸 the fitness of these +
1-x pop of defectors x Π𝒹