SlideShare ist ein Scribd-Unternehmen logo
1 von 51
Downloaden Sie, um offline zu lesen
The Minority Game:
Individual and Social Learning
under the supervision of
Dr. Matthijs Ruijgrok
Stathis Grigoropoulos
MSc Mathematics
Scientific Computing
March 27, 2014
1 / 51
Overview
1 Introduction
2 Evolutionary Game Theory (EGT)
Rationality
Learning
EGT - Nash Equilibrium
3 Revision Protocol
4 Minority Game
Stage Game
Congestion Game
Individual Learning
Social Learning
5 Conclusion
2 / 51
Introduction
• Learning has been given much attention in Artificial
Intelligence (AI) and Game Theory (GT) disciplines, as it is
the key to intelligent and rational behavior.
• What type of learning processes can we identify in a context
with many interacting agents?
• Mathematics to model learning?
• What do they learn to play? learning outcomes?
3 / 51
EGT - Rationality
• Classic Game Theory (GT) assumes perfect rationality.
• The players know all the details of the game, including each
other’s preferences over all possible outcomes.
• The richness and depth of the analysis is tremendous.
• It can handle only a small number of heterogeneous agents.
• Humans do not always play rationally
• Lack of game details, player information.
• Follow “rules of thumb”.
4 / 51
EGT - Rationality
• How do you choose the route back home to avoid
congestion?
5 / 51
EGT - Learning
• Evolutionary Game Theory (EGT), motivated by biology,
imagines the game to be repeatedly played by bounded
rational players randomly drawn from a population.
• The players make no assumption on the other players
strategies.
• Each agent has a predefined strategy of the game(like a “rule
of thumb”).
• Agents have the opportunity to adapt through an evolutionary
process over time and change their behavior.
• Can be reproduction or imitation ⇒ Learning!.
6 / 51
EGT - Learning
• Individual Learning
• In individual learning, success and failure directly influence
agent choices and behavior.
• From a AI learning perspective, individual learning is
interpreted as various types of reinforcement learning.
7 / 51
EGT - Learning
• Social Learning
• Social Learning occurs in the cases where success and failure
of other players influence choice probabilities.
• Social Learning ⇒ Social Imitation.
8 / 51
EGT - Nash Equilibrium
• Thankfully, we still have the Nash Equilibrium.
9 / 51
EGT - Nash Equilibrium
• We consider a game played by a single population, where
agents play equivalent roles.
• Let there be N players, each of whom takes a pure strategy
from the set S = {1 . . . n}.
• We call population state x the element of the simplex
X = {x ∈ Rn
+ : j∈S xj = 1}, with xj the fraction of agents
playing strategy j.
• A population game is identified by a continuous vector-valued
function that maps population states to payoffs, i.e
F : X → Rn.
• The payoff of strategy i when population state is x, is
described by the scalar Fi (x).
10 / 51
EGT - Nash Equilibrium
Population state x∗ is a Nash Equilibrium of F, when no agent can
benefit and improve his profit by switching unilaterally from
strategy i to strategy j. Specifically, x∗ is a NE if and only if:
Fi (x∗
) ≥ Fj (x∗
) ∀j ∈ S
∀ i ∈ S s.t.
x∗
i > 0.
11 / 51
Revision Protocol
• In population games introduced, agents are matched randomly
and play their strategies, producing their respected payoffs.
• However, population games can also embody cases, where all
the players take part in the game.
• The foundations of population learning dynamics are built
upon a notion called revision protocol.
• How individuals choose to update their strategy.
Definition
A revision protocol is a map ρ : Rn × X → Rn×n that takes as
input payoff vectors π and population states x and returns
non-negative matrices as output.
12 / 51
Revision Protocol - Social Learning
• Commonly, Social Imitation is modeled through the revision
protocol called proportional imitation protocol
ρij (π, x) = xj [πj − πi ]+ [7].
• Simply, imitate ‘the first man you meet on the street” with a
probability proportional to your score difference.
• This protocol generates the well-studied replicator dynamic
˙xi = xi
ˆFi (x), (1)
with ˆFi (x) = Fi (x) − ¯F(x) and ¯F(x) = i∈S xi Fi (x) [7].
13 / 51
Revision Protocol - Replicator Equation
• The replicator equation is a famous deterministic dynamic in
EGT.
• The percentage growth rate ˙xi /xi of each strategy in use, is
equal to its excess payoff.
• The attracting rest points of the replicator equation are Nash
Equilibrium (NE) of the game [9].
• Bounded Rational Agents discover NE through learning.
14 / 51
Revision Protocol - Individual Learning
• Suppose we have n populations playing an n-player game.
• A random agent from each population is selected and they
play the game at each round.
• The pairwise proportional imitation revision protocol,
• ρhk(π, xi ) = xi
k[πk − πh]+, for each population i ∈ N with
population state xi .
• We get the multi-population replicator dynamics [7].
˙xi
h = xi
h
ˆFh(xi
), ∀h ∈ S, ∀i ∈ N. (2)
15 / 51
Revision Protocol - Individual Learning
• Each population represents a player.
• Each agent within a population represents a voice or opinion
inside each the player.
• Through the imitation process the voices achieving higher
scores gets reinforced ⇒ Reinforcement Learning.
16 / 51
Minority Game
• The Minority Game is a famous Congestion Game.
• Congestion games model instances when the payoff of each
player depends on the choice of resources along with the
number of players choosing the same resource [5].
• Route Choice in road network
• Selfish packet routing in the Internet
• Congestion games are Potential Games.
• There exist a single scalar-valued function that characterizes
the game [6, p. 53].
17 / 51
Minority Game
• Consider an odd population of N agents competing in a
repeated one-shot game (N = 2k + 1, k ≥ 1) where
communication is not allowed.
• At each time step (round) t of the game, every agent has to
choose between one of two possible actions, either
αi (t) ∈ {−1, 1}.
• The minority choice wins the round at each time step and all
the winning agents are rewarded one point, none otherwise.
• By construction, the MG is a negative sum game, as the
winners are always less than the losers.
18 / 51
Minority Game
Do Individual and Social Learning lead to different learning
outcomes?
19 / 51
Minority Game - Stage Game
Definition
Define the Minority stage game as the one shot strategic game
Γ =< N, ∆, πi > consisting of:
• N players indexed by
i ∈ {1, . . . , 2k + 1}, k ∈ N, N = {1, . . . , 2k + 1},
• a finite set of strategies Ai = {−1, 1} indexed by α , where αi
denotes the strategy of player i,
• the set of mixed strategies of player i is denoted by ∆(Ai ) and
• a payoff function πi : αi × α−i → R = {0, 1}, where
α−i = i=j αj . More formally,
πi =
1 if − αi
N
j=1 αj ≥ 0
0 otherwise
(3)
20 / 51
Minority Game - Stage Game
• Some Details...
• We say agent i has mixed strategy αi :
• We think of a probability p ∈ [0, 1] that agent i plays action
{1}
• (1 − p) to play action {−1}
• Pure strategy : p = 0, 1.
• Mixed Strategy Profile α:
• All agents Mixed Strategies in a vector ⇒ Population State.
• The agents playing a mixed strategy are called mixers.
21 / 51
Minority Game - Stage Game
• “inversion” symmetry πi (−αi , −α−i ) = πi (αi , α−i )
⇒ two actions are a priori equivalent [2].
We have three general types of Nash equilibria, namely:
• Pure Strategy Nash Equilibria. I.e., when all players play a
pure strategy.
• Symmetric Mixed Strategy Nash Equilibria. That is, the
agents choose the same mixed strategy to play.
• Asymmetric Mixed Strategy Nash Equilibria. Specifically, the
NE when some players choose a pure strategy and the rest a
mixed strategy.
22 / 51
Minority Game - Stage Game
Proposition
The number of pure strategy Nash Equilibria in the stage game of
the original MG is 2 N
N−1
2
.
Proof.
The first part is the number of ways N
2 − 1 different players can be
chosen out of the set of N players at a time with minority side “0”.
Similarly, for the chosen set of players that are labeled with
winning side “1”. ♦
23 / 51
Minority Game - Stage Game
Lemma
Let be α ∈ ×i∈N∆(Ai ) a Nash equilibrium with a non-empty set of
mixers. Then all mixers use the same mixed strategy [10, 1].
Asymmetric Mixed Strategy Nash Equilibria.
• with one mixer.
• with more than one mixer.
Unique Symmetric Mixed NE
• All players choose one the two actions with probability
p = 1/2
24 / 51
Minority Game - Congestion Game
The two available choices to the agents can represent two distinct
resources in a Congestion Game.
A congestion model (N, M, (Ai )i∈N, (cj )j∈M) is described as
follows:
• N the number of players.
• M = {1..m} the number of resources.
• Ai the set of strategies of player i, where each ai ∈ Ai is a
non empty set of resources.
• For j ∈ M, cj ∈ Rn denotes the vector of benefits, where cjk is
the cost (e.g cost or payoff ) related to each user of resource
j, if there are exactly k players using that resource.
• A population state is a vector a = α = (α1, ..., α2k+1) or
point in the polyhedron ∆(A) of the mixed strategy profiles
25 / 51
Minority Game - Congestion Game
• The overall cost function for player i [8, p. 174]:
Ci =
j∈ai
cj (σj (a)) = −πi (a). (4)
Definition
A function P : A → R is a potential of the game G if
∀a ∈ A, ∀ai , bi ∈ Ai
ui (bi , a−i ) − ui (ai , a−i ) = P(bi , a−i ) − P(ai , a−i ) [3].
• Skipping a lot of nice math...
Proposition
An exact potential of the MG is the sum of the payoffs of all the
N = 2h + 1 ∈ R players. Therefore,
P(a) =
N
i=1
ui (a) ∀a ∈ A (5)
26 / 51
Minority Game - Individual Learning
• Let N = {1, ..., 2k + 1} be a set of populations, with each
population representing a Player i in the MG.
• A population state is a vector a = α = (α1, ..., α2k+1) or
point in the polyhedron ∆(A) of the mixed strategy profiles.
• Each component αi is a point in the simplex ∆(Ai ), denoting
the proportion of agents programmed to play the pure
strategy ai ∈ Ai .
• Agents – one from each population – are continuously drawn
uniformly at random from these populations to play the MG.
• Time indexed t, is continuous.
∀i ∈ N, ∀αi ∈ Ai :
˙αi (ai ) = αi (ai )(ui (ai , α−i ) − ui (αi , α−i )). (6)
27 / 51
Minority Game - Individual Learning
• A population state α ∈ ×i∈N∆(Ai ) is a stationary state of the
replicator dynamics (6),
when ˙αi (ai ) = 0 ∀i ∈ N, ∀αi ∈ Ai .
• The set T of stationary states can be partitioned into three
subsets [1].
T1 : The connected set of Nash equilibria with at most one mixer,
T2 : Nash equilibria with more than one mixer and
T3 : non-equilibrium profiles of the type (l, r, λ).



l, r ∈ 2k + 1,
l + r ≤ 2k + 1,
if l + r < 2k + 1 then λ ∈ (0, 1).
28 / 51
Minority Game - Individual Learning
• Lyapunov stability
• Asymptotic stability
29 / 51
Minority Game - Individual Learning
Concretely,
• A population state α ∈ ×i∈N∆(Ai ) is Lyapunov stable if every
neighborhood B of α contains a neighborhood B0 of α such
that ψ(t, a0) ∈ B for every x0 ∈ B ∩ ×i∈N∆(Ai ) and t ≥ 0.
• A stationary state is asymptotically stable if it is Lyapunov
stable, and, in addition, there exists a neighborhood B∗, with
limt→∞ ψ(t, a0) = α ∀α0 ∈ B∗ ∩ ×i∈N∆(Ai ).
30 / 51
Minority Game - Individual Learning
Using definition (4) of the potential function in equations (6):
∀i ∈ N, ∀αi ∈ Ai :
˙αi (ai ) = αi (ai )(U(ai , α−i ) − U(αi , α−i )). (7)
Proposition
The potential function U of the minority game is a Lyapunov
function for the replicator dynamic: for each trajectory
(α(t))t∈[0,∞], we have dU
dt ≥ 0. Equality holds at the stationary
states [1].
Proposition
The collection of Nash equilibria with at most one mixer in T1 is
asymptotically stable under the replicator dynamics. Moreover,
stationary states in T2 and T3 are not Lyapunov stable [1].
31 / 51
Minority Game - Individual Learning
• The Nash Equilibria with one mixer, are global maxima of the
potential function U.
• The potential function is the sum of the utilities of the
players. ⇒ Maximization of Utility.
• The symmetric NE of the MG is not Lyapunov stable.
32 / 51
Minority Game - Individual Learning
Three Players in the MG:
We define x, y, z ∈ R[0, 1] the probabilities for each player to play
strategy ai = 1.
A3 -1
A1/A2 -1 1
-1 000 010
1 100 001
A3 1
A1/A2 -1 1
-1 100 100
1 010 000
Table: Payoff matrix of the three-player MG. A1,A2 and A3 denote
agents 1,2,3 respectively with actions {−1, 1}. The utility matrix is split
into two submatrices using agent A3 actions as a divider. The payoffs for
each agent are presented w.r.t to their number, i.e. payoff 010 means
payoff for A1=0, A2=1 and A3=0.
33 / 51
Minority Game - Individual Learning
U(x, y, z) = u1 + u2 + u3 = x + y + z − xy − yz − zx (8)
max U(x, y, z) = 1 (9)
34 / 51
Minority Game - Individual Learning
35 / 51
Minority Game - Individual Learning
• Individual Learning is a “utilitarian” solution.
• The sum of the utilities is a Lyapunov function and is
maximized in the replicator dynamics.
• The result can be generalized to congestion games.
• What about Social Learning?
36 / 51
Minority Game - Social Learning
• No Mathematical model was attempted.
• Agent-Based Simulation were performed.
• Netlogo - Java [11],[4].
37 / 51
Minority Game - Social Learning
We define
• Attendance
Att(t) =
N
i=1
αi (t) (10)
• |Att(t)| = 1 ⇒ Nash Equilibrium!
38 / 51
Minority Game - Social Learning
Figure: The algorithm
39 / 51
Minority Game - Social Learning
Parameter Value
Agents 99,
Imitators 3
Review Rounds 3
Table: Parameters for the experiment using a Single Population of Players
40 / 51
Minority Game - Social Learning
• Not a Nash Equilibrium!
41 / 51
Minority Game - Social Learning
Figure: Histogram of strategies at stationary state,N = 99.
42 / 51
Minority Game - Social Learning
• Strategies always come as pairs.
43 / 51
Minority Game - Social Learning
44 / 51
Minority Game - Social vs Individual
Learning
45 / 51
Minority Game - Social vs Individual
Learning
46 / 51
Conclusion
• Yes! Different Learning Algorithms, do lead to different
outcomes.
• Individual Learning in MG is robust and efficient, cares about
maximization of utility.
• Individual Learning can lead to great differences in agent
performance.
• Social Learning is a more “egalitarian” solution, agents are
equal in terms of score.
• Social Learning does not converge to a Nash Equilibrium.
• Optimization ⇒ Individual Learning.
• Social Analysis ⇒ Social Learning.
47 / 51
Thank You!
Questions?
48 / 51
W. Kets and M. Voorneveld.
Congestion, Equilibrium and Learning: the Minority Game.
Discussion paper. Tilburg University, 2007.
M. Marsili, D. Challet, and R. Zecchina.
Exact solution of a modified El Farol’s bar problem: Efficiency
and the role of market impact.
Physica A Statistical Mechanics and its Applications,
280:522–553, June 2000.
Dov Monderer and Lloyd S Shapley.
Potential games.
Games and economic behavior, 14(1):124–143, 1996.
Steven F Railsback, Steven L Lytinen, and Stephen K Jackson.
Agent-based simulation platforms: Review and development
recommendations.
Simulation, 82(9):609–623, 2006.
RobertW. Rosenthal.
49 / 51
A class of games possessing pure-strategy nash equilibria.
International Journal of Game Theory, 2(1):65–67, 1973.
W.H. Sandholm.
Population Games and Evolutionary Dynamics.
Economic Learning and Social Evolution. MIT Press, 2010.
William H Sandholm.
Stochastic evolutionary game dynamics: Foundations,
deterministic approximation, and equilibrium selection.
American Mathematical Society, 201(1):1–1, 2010.
Yoav Shoham and Kevin Leyton-Brown.
Multiagent systems: Algorithmic, game-theoretic, and logical
foundations.
Cambridge University Press, 2009.
J.W. Weibull.
Evolutionary Game Theory.
MIT Press, 1997.
Duncan Whitehead.
50 / 51
The El Farol Bar Problem Revisited: Reinforcement Learning
in a Potential Game.
ESE Discussion Papers 186, Edinburgh School of Economics,
University of Edinburgh, September 2008.
U. Wilensky.
Netlogo. http://ccl.northwestern.edu/netlogo/. center for
connected learning and computer-based modeling,
northwestern university, evanston, il., 1999.
51 / 51

Weitere ähnliche Inhalte

Was ist angesagt?

Secure Hash Algorithm (SHA-512)
Secure Hash Algorithm (SHA-512)Secure Hash Algorithm (SHA-512)
Secure Hash Algorithm (SHA-512)DUET
 
Memory Management in OS
Memory Management in OSMemory Management in OS
Memory Management in OSKumar Pritam
 
Hardware and network project
Hardware and network projectHardware and network project
Hardware and network projectasif masoad
 
How DeepMind Mastered The Game Of Go
How DeepMind Mastered The Game Of GoHow DeepMind Mastered The Game Of Go
How DeepMind Mastered The Game Of GoTim Riser
 
management of distributed transactions
management of distributed transactionsmanagement of distributed transactions
management of distributed transactionsNilu Desai
 
Diffie hellman key exchange algorithm
Diffie hellman key exchange algorithmDiffie hellman key exchange algorithm
Diffie hellman key exchange algorithmSunita Kharayat
 
block ciphers
block ciphersblock ciphers
block ciphersAsad Ali
 
Deep deterministic policy gradient
Deep deterministic policy gradientDeep deterministic policy gradient
Deep deterministic policy gradientSlobodan Blazeski
 
Protection models
Protection modelsProtection models
Protection modelsG Prachi
 
Information and network security 8 security mechanisms
Information and network security 8 security mechanismsInformation and network security 8 security mechanisms
Information and network security 8 security mechanismsVaibhav Khanna
 
Travelling Salesman Problem
Travelling Salesman ProblemTravelling Salesman Problem
Travelling Salesman ProblemShikha Gupta
 
Secure Multi-Party Computation
Secure Multi-Party ComputationSecure Multi-Party Computation
Secure Multi-Party ComputationAshutosh Satapathy
 
Log based and Recovery with concurrent transaction
Log based and Recovery with concurrent transactionLog based and Recovery with concurrent transaction
Log based and Recovery with concurrent transactionnikunjandy
 

Was ist angesagt? (20)

Secure Hash Algorithm (SHA-512)
Secure Hash Algorithm (SHA-512)Secure Hash Algorithm (SHA-512)
Secure Hash Algorithm (SHA-512)
 
Disjoint sets
Disjoint setsDisjoint sets
Disjoint sets
 
Memory Management in OS
Memory Management in OSMemory Management in OS
Memory Management in OS
 
Hardware and network project
Hardware and network projectHardware and network project
Hardware and network project
 
How DeepMind Mastered The Game Of Go
How DeepMind Mastered The Game Of GoHow DeepMind Mastered The Game Of Go
How DeepMind Mastered The Game Of Go
 
management of distributed transactions
management of distributed transactionsmanagement of distributed transactions
management of distributed transactions
 
Recursive algorithms
Recursive algorithmsRecursive algorithms
Recursive algorithms
 
Diffie hellman key exchange algorithm
Diffie hellman key exchange algorithmDiffie hellman key exchange algorithm
Diffie hellman key exchange algorithm
 
block ciphers
block ciphersblock ciphers
block ciphers
 
Random forest
Random forestRandom forest
Random forest
 
Neural network
Neural networkNeural network
Neural network
 
Deep deterministic policy gradient
Deep deterministic policy gradientDeep deterministic policy gradient
Deep deterministic policy gradient
 
Protection models
Protection modelsProtection models
Protection models
 
Information and network security 8 security mechanisms
Information and network security 8 security mechanismsInformation and network security 8 security mechanisms
Information and network security 8 security mechanisms
 
Unit 3 daa
Unit 3 daaUnit 3 daa
Unit 3 daa
 
Travelling Salesman Problem
Travelling Salesman ProblemTravelling Salesman Problem
Travelling Salesman Problem
 
Fuzzy Clustering(C-means, K-means)
Fuzzy Clustering(C-means, K-means)Fuzzy Clustering(C-means, K-means)
Fuzzy Clustering(C-means, K-means)
 
Secure Multi-Party Computation
Secure Multi-Party ComputationSecure Multi-Party Computation
Secure Multi-Party Computation
 
Log based and Recovery with concurrent transaction
Log based and Recovery with concurrent transactionLog based and Recovery with concurrent transaction
Log based and Recovery with concurrent transaction
 
Associative memory network
Associative memory networkAssociative memory network
Associative memory network
 

Andere mochten auch

Research remix social_learning
Research remix social_learningResearch remix social_learning
Research remix social_learningShannon Parker
 
Skylark Esta Brochure - Zricks.com
Skylark Esta Brochure - Zricks.comSkylark Esta Brochure - Zricks.com
Skylark Esta Brochure - Zricks.comZricks.com
 
Viadotto Paralà: scheda opera d'arte di ingegneria strutturale redatta nell' ...
Viadotto Paralà: scheda opera d'arte di ingegneria strutturale redatta nell' ...Viadotto Paralà: scheda opera d'arte di ingegneria strutturale redatta nell' ...
Viadotto Paralà: scheda opera d'arte di ingegneria strutturale redatta nell' ...Albina Bianca Maria Cotza
 
FiNPLACE 2016. Размер имеет значение. Shukhrat Yakubov Google
FiNPLACE 2016. Размер имеет значение. Shukhrat Yakubov GoogleFiNPLACE 2016. Размер имеет значение. Shukhrat Yakubov Google
FiNPLACE 2016. Размер имеет значение. Shukhrat Yakubov GoogleShukhrat Yakubov
 
Daughters Without Dads Inc
Daughters Without Dads IncDaughters Without Dads Inc
Daughters Without Dads Incarmstrongdoresa
 
ประเภทของโครงงานคอมพิวเตอร์
ประเภทของโครงงานคอมพิวเตอร์ประเภทของโครงงานคอมพิวเตอร์
ประเภทของโครงงานคอมพิวเตอร์122 Chen
 
Mother teresa of calcutta copia
Mother   teresa   of   calcutta   copiaMother   teresa   of   calcutta   copia
Mother teresa of calcutta copialupatric8
 
American freedom defense_initiative_group_2_final paper
American freedom defense_initiative_group_2_final paperAmerican freedom defense_initiative_group_2_final paper
American freedom defense_initiative_group_2_final paperUna Hrnjak
 
My dream comes a nightmare
My dream comes a nightmareMy dream comes a nightmare
My dream comes a nightmarelanguagefy
 
The Supply of Sustainable Tourism and The Potential Challenges Associated Wit...
The Supply of Sustainable Tourism and The Potential Challenges Associated Wit...The Supply of Sustainable Tourism and The Potential Challenges Associated Wit...
The Supply of Sustainable Tourism and The Potential Challenges Associated Wit...Aine Conaghan
 

Andere mochten auch (20)

How to apply to 1 year programs
How to apply to 1 year programsHow to apply to 1 year programs
How to apply to 1 year programs
 
Research remix social_learning
Research remix social_learningResearch remix social_learning
Research remix social_learning
 
HR and Efficency Solutions
HR and Efficency SolutionsHR and Efficency Solutions
HR and Efficency Solutions
 
Skylark Esta Brochure - Zricks.com
Skylark Esta Brochure - Zricks.comSkylark Esta Brochure - Zricks.com
Skylark Esta Brochure - Zricks.com
 
Grade 4
Grade 4Grade 4
Grade 4
 
3 sun powered
3 sun powered3 sun powered
3 sun powered
 
Viadotto Paralà: scheda opera d'arte di ingegneria strutturale redatta nell' ...
Viadotto Paralà: scheda opera d'arte di ingegneria strutturale redatta nell' ...Viadotto Paralà: scheda opera d'arte di ingegneria strutturale redatta nell' ...
Viadotto Paralà: scheda opera d'arte di ingegneria strutturale redatta nell' ...
 
FiNPLACE 2016. Размер имеет значение. Shukhrat Yakubov Google
FiNPLACE 2016. Размер имеет значение. Shukhrat Yakubov GoogleFiNPLACE 2016. Размер имеет значение. Shukhrat Yakubov Google
FiNPLACE 2016. Размер имеет значение. Shukhrat Yakubov Google
 
Daughters Without Dads Inc
Daughters Without Dads IncDaughters Without Dads Inc
Daughters Without Dads Inc
 
ประเภทของโครงงานคอมพิวเตอร์
ประเภทของโครงงานคอมพิวเตอร์ประเภทของโครงงานคอมพิวเตอร์
ประเภทของโครงงานคอมพิวเตอร์
 
Mother teresa of calcutta copia
Mother   teresa   of   calcutta   copiaMother   teresa   of   calcutta   copia
Mother teresa of calcutta copia
 
Estudo para Naming
Estudo para Naming Estudo para Naming
Estudo para Naming
 
American freedom defense_initiative_group_2_final paper
American freedom defense_initiative_group_2_final paperAmerican freedom defense_initiative_group_2_final paper
American freedom defense_initiative_group_2_final paper
 
My dream comes a nightmare
My dream comes a nightmareMy dream comes a nightmare
My dream comes a nightmare
 
Presentation5,000
Presentation5,000Presentation5,000
Presentation5,000
 
Hippa privacy
Hippa privacyHippa privacy
Hippa privacy
 
Bbb
BbbBbb
Bbb
 
The Supply of Sustainable Tourism and The Potential Challenges Associated Wit...
The Supply of Sustainable Tourism and The Potential Challenges Associated Wit...The Supply of Sustainable Tourism and The Potential Challenges Associated Wit...
The Supply of Sustainable Tourism and The Potential Challenges Associated Wit...
 
Debugging
DebuggingDebugging
Debugging
 
Negotiating the offer
Negotiating the offerNegotiating the offer
Negotiating the offer
 

Ähnlich wie The Minority Game: Individual and Social Learning

Ähnlich wie The Minority Game: Individual and Social Learning (20)

Dynamics
DynamicsDynamics
Dynamics
 
file1
file1file1
file1
 
Game Theory in Cryptoeconomics
Game Theory in CryptoeconomicsGame Theory in Cryptoeconomics
Game Theory in Cryptoeconomics
 
AI3391 Artificial Intelligence Session 14 Adversarial Search .pptx
AI3391 Artificial Intelligence Session 14 Adversarial Search .pptxAI3391 Artificial Intelligence Session 14 Adversarial Search .pptx
AI3391 Artificial Intelligence Session 14 Adversarial Search .pptx
 
Module 3 Game Theory (1).pptx
Module 3 Game Theory (1).pptxModule 3 Game Theory (1).pptx
Module 3 Game Theory (1).pptx
 
OR PPT 280322 maximin final - nikhil tiwari.pptx
OR PPT 280322 maximin final - nikhil tiwari.pptxOR PPT 280322 maximin final - nikhil tiwari.pptx
OR PPT 280322 maximin final - nikhil tiwari.pptx
 
Game Theory SV.docx
Game Theory SV.docxGame Theory SV.docx
Game Theory SV.docx
 
TermPaper
TermPaperTermPaper
TermPaper
 
Game theory
Game theoryGame theory
Game theory
 
cvpr2011: game theory in CVPR part 1
cvpr2011: game theory in CVPR part 1cvpr2011: game theory in CVPR part 1
cvpr2011: game theory in CVPR part 1
 
An introduction to Game Theory
An introduction to Game TheoryAn introduction to Game Theory
An introduction to Game Theory
 
present_2_new
present_2_newpresent_2_new
present_2_new
 
L6
L6L6
L6
 
Dissertation Conference Poster
Dissertation Conference PosterDissertation Conference Poster
Dissertation Conference Poster
 
Optimization of Fuzzy Matrix Games of Order 4 X 3
Optimization of Fuzzy Matrix Games of Order 4 X 3Optimization of Fuzzy Matrix Games of Order 4 X 3
Optimization of Fuzzy Matrix Games of Order 4 X 3
 
Game theory
Game theoryGame theory
Game theory
 
Game Theory Economics
Game Theory EconomicsGame Theory Economics
Game Theory Economics
 
Lect04 slides
Lect04 slidesLect04 slides
Lect04 slides
 
CAGT-IST Student Presentations
CAGT-IST Student Presentations CAGT-IST Student Presentations
CAGT-IST Student Presentations
 
navingameppt-191018085333.pdf
navingameppt-191018085333.pdfnavingameppt-191018085333.pdf
navingameppt-191018085333.pdf
 

Mehr von Stathis Grigoropoulos

Mehr von Stathis Grigoropoulos (8)

Agent Based Modeling and Simulation - Overview and Tools
Agent Based Modeling and Simulation - Overview and ToolsAgent Based Modeling and Simulation - Overview and Tools
Agent Based Modeling and Simulation - Overview and Tools
 
Social cognitive factors of unfair ratings in reputation
Social cognitive factors of unfair ratings in reputationSocial cognitive factors of unfair ratings in reputation
Social cognitive factors of unfair ratings in reputation
 
Mechanism design theory examples and complexity
Mechanism design theory examples and complexityMechanism design theory examples and complexity
Mechanism design theory examples and complexity
 
Chess Segmentation
Chess SegmentationChess Segmentation
Chess Segmentation
 
Scheduling
SchedulingScheduling
Scheduling
 
Scheduling
SchedulingScheduling
Scheduling
 
Hydrate me
Hydrate meHydrate me
Hydrate me
 
Stress2
Stress2Stress2
Stress2
 

Kürzlich hochgeladen

Explainable AI for distinguishing future climate change scenarios
Explainable AI for distinguishing future climate change scenariosExplainable AI for distinguishing future climate change scenarios
Explainable AI for distinguishing future climate change scenariosZachary Labe
 
Observational constraints on mergers creating magnetism in massive stars
Observational constraints on mergers creating magnetism in massive starsObservational constraints on mergers creating magnetism in massive stars
Observational constraints on mergers creating magnetism in massive starsSérgio Sacani
 
Gas-ExchangeS-in-Plants-and-Animals.pptx
Gas-ExchangeS-in-Plants-and-Animals.pptxGas-ExchangeS-in-Plants-and-Animals.pptx
Gas-ExchangeS-in-Plants-and-Animals.pptxGiovaniTrinidad
 
complex analysis best book for solving questions.pdf
complex analysis best book for solving questions.pdfcomplex analysis best book for solving questions.pdf
complex analysis best book for solving questions.pdfSubhamKumar3239
 
6.1 Pests of Groundnut_Binomics_Identification_Dr.UPR
6.1 Pests of Groundnut_Binomics_Identification_Dr.UPR6.1 Pests of Groundnut_Binomics_Identification_Dr.UPR
6.1 Pests of Groundnut_Binomics_Identification_Dr.UPRPirithiRaju
 
whole genome sequencing new and its types including shortgun and clone by clone
whole genome sequencing new  and its types including shortgun and clone by clonewhole genome sequencing new  and its types including shortgun and clone by clone
whole genome sequencing new and its types including shortgun and clone by clonechaudhary charan shingh university
 
Quarter 4_Grade 8_Digestive System Structure and Functions
Quarter 4_Grade 8_Digestive System Structure and FunctionsQuarter 4_Grade 8_Digestive System Structure and Functions
Quarter 4_Grade 8_Digestive System Structure and FunctionsCharlene Llagas
 
Fertilization: Sperm and the egg—collectively called the gametes—fuse togethe...
Fertilization: Sperm and the egg—collectively called the gametes—fuse togethe...Fertilization: Sperm and the egg—collectively called the gametes—fuse togethe...
Fertilization: Sperm and the egg—collectively called the gametes—fuse togethe...D. B. S. College Kanpur
 
办理麦克马斯特大学毕业证成绩单|购买加拿大文凭证书
办理麦克马斯特大学毕业证成绩单|购买加拿大文凭证书办理麦克马斯特大学毕业证成绩单|购买加拿大文凭证书
办理麦克马斯特大学毕业证成绩单|购买加拿大文凭证书zdzoqco
 
Environmental acoustics- noise criteria.pptx
Environmental acoustics- noise criteria.pptxEnvironmental acoustics- noise criteria.pptx
Environmental acoustics- noise criteria.pptxpriyankatabhane
 
Pests of Sunflower_Binomics_Identification_Dr.UPR
Pests of Sunflower_Binomics_Identification_Dr.UPRPests of Sunflower_Binomics_Identification_Dr.UPR
Pests of Sunflower_Binomics_Identification_Dr.UPRPirithiRaju
 
final waves properties grade 7 - third quarter
final waves properties grade 7 - third quarterfinal waves properties grade 7 - third quarter
final waves properties grade 7 - third quarterHanHyoKim
 
GLYCOSIDES Classification Of GLYCOSIDES Chemical Tests Glycosides
GLYCOSIDES Classification Of GLYCOSIDES  Chemical Tests GlycosidesGLYCOSIDES Classification Of GLYCOSIDES  Chemical Tests Glycosides
GLYCOSIDES Classification Of GLYCOSIDES Chemical Tests GlycosidesNandakishor Bhaurao Deshmukh
 
GenAI talk for Young at Wageningen University & Research (WUR) March 2024
GenAI talk for Young at Wageningen University & Research (WUR) March 2024GenAI talk for Young at Wageningen University & Research (WUR) March 2024
GenAI talk for Young at Wageningen University & Research (WUR) March 2024Jene van der Heide
 
How we decide powerpoint presentation.pptx
How we decide powerpoint presentation.pptxHow we decide powerpoint presentation.pptx
How we decide powerpoint presentation.pptxJosielynTars
 
linear Regression, multiple Regression and Annova
linear Regression, multiple Regression and Annovalinear Regression, multiple Regression and Annova
linear Regression, multiple Regression and AnnovaMansi Rastogi
 
Environmental Acoustics- Speech interference level, acoustics calibrator.pptx
Environmental Acoustics- Speech interference level, acoustics calibrator.pptxEnvironmental Acoustics- Speech interference level, acoustics calibrator.pptx
Environmental Acoustics- Speech interference level, acoustics calibrator.pptxpriyankatabhane
 
Introduction of Human Body & Structure of cell.pptx
Introduction of Human Body & Structure of cell.pptxIntroduction of Human Body & Structure of cell.pptx
Introduction of Human Body & Structure of cell.pptxMedical College
 

Kürzlich hochgeladen (20)

Explainable AI for distinguishing future climate change scenarios
Explainable AI for distinguishing future climate change scenariosExplainable AI for distinguishing future climate change scenarios
Explainable AI for distinguishing future climate change scenarios
 
Observational constraints on mergers creating magnetism in massive stars
Observational constraints on mergers creating magnetism in massive starsObservational constraints on mergers creating magnetism in massive stars
Observational constraints on mergers creating magnetism in massive stars
 
Interferons.pptx.
Interferons.pptx.Interferons.pptx.
Interferons.pptx.
 
Gas-ExchangeS-in-Plants-and-Animals.pptx
Gas-ExchangeS-in-Plants-and-Animals.pptxGas-ExchangeS-in-Plants-and-Animals.pptx
Gas-ExchangeS-in-Plants-and-Animals.pptx
 
complex analysis best book for solving questions.pdf
complex analysis best book for solving questions.pdfcomplex analysis best book for solving questions.pdf
complex analysis best book for solving questions.pdf
 
6.1 Pests of Groundnut_Binomics_Identification_Dr.UPR
6.1 Pests of Groundnut_Binomics_Identification_Dr.UPR6.1 Pests of Groundnut_Binomics_Identification_Dr.UPR
6.1 Pests of Groundnut_Binomics_Identification_Dr.UPR
 
whole genome sequencing new and its types including shortgun and clone by clone
whole genome sequencing new  and its types including shortgun and clone by clonewhole genome sequencing new  and its types including shortgun and clone by clone
whole genome sequencing new and its types including shortgun and clone by clone
 
Quarter 4_Grade 8_Digestive System Structure and Functions
Quarter 4_Grade 8_Digestive System Structure and FunctionsQuarter 4_Grade 8_Digestive System Structure and Functions
Quarter 4_Grade 8_Digestive System Structure and Functions
 
Fertilization: Sperm and the egg—collectively called the gametes—fuse togethe...
Fertilization: Sperm and the egg—collectively called the gametes—fuse togethe...Fertilization: Sperm and the egg—collectively called the gametes—fuse togethe...
Fertilization: Sperm and the egg—collectively called the gametes—fuse togethe...
 
办理麦克马斯特大学毕业证成绩单|购买加拿大文凭证书
办理麦克马斯特大学毕业证成绩单|购买加拿大文凭证书办理麦克马斯特大学毕业证成绩单|购买加拿大文凭证书
办理麦克马斯特大学毕业证成绩单|购买加拿大文凭证书
 
Environmental acoustics- noise criteria.pptx
Environmental acoustics- noise criteria.pptxEnvironmental acoustics- noise criteria.pptx
Environmental acoustics- noise criteria.pptx
 
Pests of Sunflower_Binomics_Identification_Dr.UPR
Pests of Sunflower_Binomics_Identification_Dr.UPRPests of Sunflower_Binomics_Identification_Dr.UPR
Pests of Sunflower_Binomics_Identification_Dr.UPR
 
final waves properties grade 7 - third quarter
final waves properties grade 7 - third quarterfinal waves properties grade 7 - third quarter
final waves properties grade 7 - third quarter
 
GLYCOSIDES Classification Of GLYCOSIDES Chemical Tests Glycosides
GLYCOSIDES Classification Of GLYCOSIDES  Chemical Tests GlycosidesGLYCOSIDES Classification Of GLYCOSIDES  Chemical Tests Glycosides
GLYCOSIDES Classification Of GLYCOSIDES Chemical Tests Glycosides
 
GenAI talk for Young at Wageningen University & Research (WUR) March 2024
GenAI talk for Young at Wageningen University & Research (WUR) March 2024GenAI talk for Young at Wageningen University & Research (WUR) March 2024
GenAI talk for Young at Wageningen University & Research (WUR) March 2024
 
How we decide powerpoint presentation.pptx
How we decide powerpoint presentation.pptxHow we decide powerpoint presentation.pptx
How we decide powerpoint presentation.pptx
 
linear Regression, multiple Regression and Annova
linear Regression, multiple Regression and Annovalinear Regression, multiple Regression and Annova
linear Regression, multiple Regression and Annova
 
Environmental Acoustics- Speech interference level, acoustics calibrator.pptx
Environmental Acoustics- Speech interference level, acoustics calibrator.pptxEnvironmental Acoustics- Speech interference level, acoustics calibrator.pptx
Environmental Acoustics- Speech interference level, acoustics calibrator.pptx
 
Introduction of Human Body & Structure of cell.pptx
Introduction of Human Body & Structure of cell.pptxIntroduction of Human Body & Structure of cell.pptx
Introduction of Human Body & Structure of cell.pptx
 
AZOTOBACTER AS BIOFERILIZER.PPTX
AZOTOBACTER AS BIOFERILIZER.PPTXAZOTOBACTER AS BIOFERILIZER.PPTX
AZOTOBACTER AS BIOFERILIZER.PPTX
 

The Minority Game: Individual and Social Learning

  • 1. The Minority Game: Individual and Social Learning under the supervision of Dr. Matthijs Ruijgrok Stathis Grigoropoulos MSc Mathematics Scientific Computing March 27, 2014 1 / 51
  • 2. Overview 1 Introduction 2 Evolutionary Game Theory (EGT) Rationality Learning EGT - Nash Equilibrium 3 Revision Protocol 4 Minority Game Stage Game Congestion Game Individual Learning Social Learning 5 Conclusion 2 / 51
  • 3. Introduction • Learning has been given much attention in Artificial Intelligence (AI) and Game Theory (GT) disciplines, as it is the key to intelligent and rational behavior. • What type of learning processes can we identify in a context with many interacting agents? • Mathematics to model learning? • What do they learn to play? learning outcomes? 3 / 51
  • 4. EGT - Rationality • Classic Game Theory (GT) assumes perfect rationality. • The players know all the details of the game, including each other’s preferences over all possible outcomes. • The richness and depth of the analysis is tremendous. • It can handle only a small number of heterogeneous agents. • Humans do not always play rationally • Lack of game details, player information. • Follow “rules of thumb”. 4 / 51
  • 5. EGT - Rationality • How do you choose the route back home to avoid congestion? 5 / 51
  • 6. EGT - Learning • Evolutionary Game Theory (EGT), motivated by biology, imagines the game to be repeatedly played by bounded rational players randomly drawn from a population. • The players make no assumption on the other players strategies. • Each agent has a predefined strategy of the game(like a “rule of thumb”). • Agents have the opportunity to adapt through an evolutionary process over time and change their behavior. • Can be reproduction or imitation ⇒ Learning!. 6 / 51
  • 7. EGT - Learning • Individual Learning • In individual learning, success and failure directly influence agent choices and behavior. • From a AI learning perspective, individual learning is interpreted as various types of reinforcement learning. 7 / 51
  • 8. EGT - Learning • Social Learning • Social Learning occurs in the cases where success and failure of other players influence choice probabilities. • Social Learning ⇒ Social Imitation. 8 / 51
  • 9. EGT - Nash Equilibrium • Thankfully, we still have the Nash Equilibrium. 9 / 51
  • 10. EGT - Nash Equilibrium • We consider a game played by a single population, where agents play equivalent roles. • Let there be N players, each of whom takes a pure strategy from the set S = {1 . . . n}. • We call population state x the element of the simplex X = {x ∈ Rn + : j∈S xj = 1}, with xj the fraction of agents playing strategy j. • A population game is identified by a continuous vector-valued function that maps population states to payoffs, i.e F : X → Rn. • The payoff of strategy i when population state is x, is described by the scalar Fi (x). 10 / 51
  • 11. EGT - Nash Equilibrium Population state x∗ is a Nash Equilibrium of F, when no agent can benefit and improve his profit by switching unilaterally from strategy i to strategy j. Specifically, x∗ is a NE if and only if: Fi (x∗ ) ≥ Fj (x∗ ) ∀j ∈ S ∀ i ∈ S s.t. x∗ i > 0. 11 / 51
  • 12. Revision Protocol • In population games introduced, agents are matched randomly and play their strategies, producing their respected payoffs. • However, population games can also embody cases, where all the players take part in the game. • The foundations of population learning dynamics are built upon a notion called revision protocol. • How individuals choose to update their strategy. Definition A revision protocol is a map ρ : Rn × X → Rn×n that takes as input payoff vectors π and population states x and returns non-negative matrices as output. 12 / 51
  • 13. Revision Protocol - Social Learning • Commonly, Social Imitation is modeled through the revision protocol called proportional imitation protocol ρij (π, x) = xj [πj − πi ]+ [7]. • Simply, imitate ‘the first man you meet on the street” with a probability proportional to your score difference. • This protocol generates the well-studied replicator dynamic ˙xi = xi ˆFi (x), (1) with ˆFi (x) = Fi (x) − ¯F(x) and ¯F(x) = i∈S xi Fi (x) [7]. 13 / 51
  • 14. Revision Protocol - Replicator Equation • The replicator equation is a famous deterministic dynamic in EGT. • The percentage growth rate ˙xi /xi of each strategy in use, is equal to its excess payoff. • The attracting rest points of the replicator equation are Nash Equilibrium (NE) of the game [9]. • Bounded Rational Agents discover NE through learning. 14 / 51
  • 15. Revision Protocol - Individual Learning • Suppose we have n populations playing an n-player game. • A random agent from each population is selected and they play the game at each round. • The pairwise proportional imitation revision protocol, • ρhk(π, xi ) = xi k[πk − πh]+, for each population i ∈ N with population state xi . • We get the multi-population replicator dynamics [7]. ˙xi h = xi h ˆFh(xi ), ∀h ∈ S, ∀i ∈ N. (2) 15 / 51
  • 16. Revision Protocol - Individual Learning • Each population represents a player. • Each agent within a population represents a voice or opinion inside each the player. • Through the imitation process the voices achieving higher scores gets reinforced ⇒ Reinforcement Learning. 16 / 51
  • 17. Minority Game • The Minority Game is a famous Congestion Game. • Congestion games model instances when the payoff of each player depends on the choice of resources along with the number of players choosing the same resource [5]. • Route Choice in road network • Selfish packet routing in the Internet • Congestion games are Potential Games. • There exist a single scalar-valued function that characterizes the game [6, p. 53]. 17 / 51
  • 18. Minority Game • Consider an odd population of N agents competing in a repeated one-shot game (N = 2k + 1, k ≥ 1) where communication is not allowed. • At each time step (round) t of the game, every agent has to choose between one of two possible actions, either αi (t) ∈ {−1, 1}. • The minority choice wins the round at each time step and all the winning agents are rewarded one point, none otherwise. • By construction, the MG is a negative sum game, as the winners are always less than the losers. 18 / 51
  • 19. Minority Game Do Individual and Social Learning lead to different learning outcomes? 19 / 51
  • 20. Minority Game - Stage Game Definition Define the Minority stage game as the one shot strategic game Γ =< N, ∆, πi > consisting of: • N players indexed by i ∈ {1, . . . , 2k + 1}, k ∈ N, N = {1, . . . , 2k + 1}, • a finite set of strategies Ai = {−1, 1} indexed by α , where αi denotes the strategy of player i, • the set of mixed strategies of player i is denoted by ∆(Ai ) and • a payoff function πi : αi × α−i → R = {0, 1}, where α−i = i=j αj . More formally, πi = 1 if − αi N j=1 αj ≥ 0 0 otherwise (3) 20 / 51
  • 21. Minority Game - Stage Game • Some Details... • We say agent i has mixed strategy αi : • We think of a probability p ∈ [0, 1] that agent i plays action {1} • (1 − p) to play action {−1} • Pure strategy : p = 0, 1. • Mixed Strategy Profile α: • All agents Mixed Strategies in a vector ⇒ Population State. • The agents playing a mixed strategy are called mixers. 21 / 51
  • 22. Minority Game - Stage Game • “inversion” symmetry πi (−αi , −α−i ) = πi (αi , α−i ) ⇒ two actions are a priori equivalent [2]. We have three general types of Nash equilibria, namely: • Pure Strategy Nash Equilibria. I.e., when all players play a pure strategy. • Symmetric Mixed Strategy Nash Equilibria. That is, the agents choose the same mixed strategy to play. • Asymmetric Mixed Strategy Nash Equilibria. Specifically, the NE when some players choose a pure strategy and the rest a mixed strategy. 22 / 51
  • 23. Minority Game - Stage Game Proposition The number of pure strategy Nash Equilibria in the stage game of the original MG is 2 N N−1 2 . Proof. The first part is the number of ways N 2 − 1 different players can be chosen out of the set of N players at a time with minority side “0”. Similarly, for the chosen set of players that are labeled with winning side “1”. ♦ 23 / 51
  • 24. Minority Game - Stage Game Lemma Let be α ∈ ×i∈N∆(Ai ) a Nash equilibrium with a non-empty set of mixers. Then all mixers use the same mixed strategy [10, 1]. Asymmetric Mixed Strategy Nash Equilibria. • with one mixer. • with more than one mixer. Unique Symmetric Mixed NE • All players choose one the two actions with probability p = 1/2 24 / 51
  • 25. Minority Game - Congestion Game The two available choices to the agents can represent two distinct resources in a Congestion Game. A congestion model (N, M, (Ai )i∈N, (cj )j∈M) is described as follows: • N the number of players. • M = {1..m} the number of resources. • Ai the set of strategies of player i, where each ai ∈ Ai is a non empty set of resources. • For j ∈ M, cj ∈ Rn denotes the vector of benefits, where cjk is the cost (e.g cost or payoff ) related to each user of resource j, if there are exactly k players using that resource. • A population state is a vector a = α = (α1, ..., α2k+1) or point in the polyhedron ∆(A) of the mixed strategy profiles 25 / 51
  • 26. Minority Game - Congestion Game • The overall cost function for player i [8, p. 174]: Ci = j∈ai cj (σj (a)) = −πi (a). (4) Definition A function P : A → R is a potential of the game G if ∀a ∈ A, ∀ai , bi ∈ Ai ui (bi , a−i ) − ui (ai , a−i ) = P(bi , a−i ) − P(ai , a−i ) [3]. • Skipping a lot of nice math... Proposition An exact potential of the MG is the sum of the payoffs of all the N = 2h + 1 ∈ R players. Therefore, P(a) = N i=1 ui (a) ∀a ∈ A (5) 26 / 51
  • 27. Minority Game - Individual Learning • Let N = {1, ..., 2k + 1} be a set of populations, with each population representing a Player i in the MG. • A population state is a vector a = α = (α1, ..., α2k+1) or point in the polyhedron ∆(A) of the mixed strategy profiles. • Each component αi is a point in the simplex ∆(Ai ), denoting the proportion of agents programmed to play the pure strategy ai ∈ Ai . • Agents – one from each population – are continuously drawn uniformly at random from these populations to play the MG. • Time indexed t, is continuous. ∀i ∈ N, ∀αi ∈ Ai : ˙αi (ai ) = αi (ai )(ui (ai , α−i ) − ui (αi , α−i )). (6) 27 / 51
  • 28. Minority Game - Individual Learning • A population state α ∈ ×i∈N∆(Ai ) is a stationary state of the replicator dynamics (6), when ˙αi (ai ) = 0 ∀i ∈ N, ∀αi ∈ Ai . • The set T of stationary states can be partitioned into three subsets [1]. T1 : The connected set of Nash equilibria with at most one mixer, T2 : Nash equilibria with more than one mixer and T3 : non-equilibrium profiles of the type (l, r, λ).    l, r ∈ 2k + 1, l + r ≤ 2k + 1, if l + r < 2k + 1 then λ ∈ (0, 1). 28 / 51
  • 29. Minority Game - Individual Learning • Lyapunov stability • Asymptotic stability 29 / 51
  • 30. Minority Game - Individual Learning Concretely, • A population state α ∈ ×i∈N∆(Ai ) is Lyapunov stable if every neighborhood B of α contains a neighborhood B0 of α such that ψ(t, a0) ∈ B for every x0 ∈ B ∩ ×i∈N∆(Ai ) and t ≥ 0. • A stationary state is asymptotically stable if it is Lyapunov stable, and, in addition, there exists a neighborhood B∗, with limt→∞ ψ(t, a0) = α ∀α0 ∈ B∗ ∩ ×i∈N∆(Ai ). 30 / 51
  • 31. Minority Game - Individual Learning Using definition (4) of the potential function in equations (6): ∀i ∈ N, ∀αi ∈ Ai : ˙αi (ai ) = αi (ai )(U(ai , α−i ) − U(αi , α−i )). (7) Proposition The potential function U of the minority game is a Lyapunov function for the replicator dynamic: for each trajectory (α(t))t∈[0,∞], we have dU dt ≥ 0. Equality holds at the stationary states [1]. Proposition The collection of Nash equilibria with at most one mixer in T1 is asymptotically stable under the replicator dynamics. Moreover, stationary states in T2 and T3 are not Lyapunov stable [1]. 31 / 51
  • 32. Minority Game - Individual Learning • The Nash Equilibria with one mixer, are global maxima of the potential function U. • The potential function is the sum of the utilities of the players. ⇒ Maximization of Utility. • The symmetric NE of the MG is not Lyapunov stable. 32 / 51
  • 33. Minority Game - Individual Learning Three Players in the MG: We define x, y, z ∈ R[0, 1] the probabilities for each player to play strategy ai = 1. A3 -1 A1/A2 -1 1 -1 000 010 1 100 001 A3 1 A1/A2 -1 1 -1 100 100 1 010 000 Table: Payoff matrix of the three-player MG. A1,A2 and A3 denote agents 1,2,3 respectively with actions {−1, 1}. The utility matrix is split into two submatrices using agent A3 actions as a divider. The payoffs for each agent are presented w.r.t to their number, i.e. payoff 010 means payoff for A1=0, A2=1 and A3=0. 33 / 51
  • 34. Minority Game - Individual Learning U(x, y, z) = u1 + u2 + u3 = x + y + z − xy − yz − zx (8) max U(x, y, z) = 1 (9) 34 / 51
  • 35. Minority Game - Individual Learning 35 / 51
  • 36. Minority Game - Individual Learning • Individual Learning is a “utilitarian” solution. • The sum of the utilities is a Lyapunov function and is maximized in the replicator dynamics. • The result can be generalized to congestion games. • What about Social Learning? 36 / 51
  • 37. Minority Game - Social Learning • No Mathematical model was attempted. • Agent-Based Simulation were performed. • Netlogo - Java [11],[4]. 37 / 51
  • 38. Minority Game - Social Learning We define • Attendance Att(t) = N i=1 αi (t) (10) • |Att(t)| = 1 ⇒ Nash Equilibrium! 38 / 51
  • 39. Minority Game - Social Learning Figure: The algorithm 39 / 51
  • 40. Minority Game - Social Learning Parameter Value Agents 99, Imitators 3 Review Rounds 3 Table: Parameters for the experiment using a Single Population of Players 40 / 51
  • 41. Minority Game - Social Learning • Not a Nash Equilibrium! 41 / 51
  • 42. Minority Game - Social Learning Figure: Histogram of strategies at stationary state,N = 99. 42 / 51
  • 43. Minority Game - Social Learning • Strategies always come as pairs. 43 / 51
  • 44. Minority Game - Social Learning 44 / 51
  • 45. Minority Game - Social vs Individual Learning 45 / 51
  • 46. Minority Game - Social vs Individual Learning 46 / 51
  • 47. Conclusion • Yes! Different Learning Algorithms, do lead to different outcomes. • Individual Learning in MG is robust and efficient, cares about maximization of utility. • Individual Learning can lead to great differences in agent performance. • Social Learning is a more “egalitarian” solution, agents are equal in terms of score. • Social Learning does not converge to a Nash Equilibrium. • Optimization ⇒ Individual Learning. • Social Analysis ⇒ Social Learning. 47 / 51
  • 49. W. Kets and M. Voorneveld. Congestion, Equilibrium and Learning: the Minority Game. Discussion paper. Tilburg University, 2007. M. Marsili, D. Challet, and R. Zecchina. Exact solution of a modified El Farol’s bar problem: Efficiency and the role of market impact. Physica A Statistical Mechanics and its Applications, 280:522–553, June 2000. Dov Monderer and Lloyd S Shapley. Potential games. Games and economic behavior, 14(1):124–143, 1996. Steven F Railsback, Steven L Lytinen, and Stephen K Jackson. Agent-based simulation platforms: Review and development recommendations. Simulation, 82(9):609–623, 2006. RobertW. Rosenthal. 49 / 51
  • 50. A class of games possessing pure-strategy nash equilibria. International Journal of Game Theory, 2(1):65–67, 1973. W.H. Sandholm. Population Games and Evolutionary Dynamics. Economic Learning and Social Evolution. MIT Press, 2010. William H Sandholm. Stochastic evolutionary game dynamics: Foundations, deterministic approximation, and equilibrium selection. American Mathematical Society, 201(1):1–1, 2010. Yoav Shoham and Kevin Leyton-Brown. Multiagent systems: Algorithmic, game-theoretic, and logical foundations. Cambridge University Press, 2009. J.W. Weibull. Evolutionary Game Theory. MIT Press, 1997. Duncan Whitehead. 50 / 51
  • 51. The El Farol Bar Problem Revisited: Reinforcement Learning in a Potential Game. ESE Discussion Papers 186, Edinburgh School of Economics, University of Edinburgh, September 2008. U. Wilensky. Netlogo. http://ccl.northwestern.edu/netlogo/. center for connected learning and computer-based modeling, northwestern university, evanston, il., 1999. 51 / 51