2024 Game Theory
2024 Game Theory
Education background:
2008.09-2012.06, Guangxi University, Bachelor
2012.09-2015.06, Northwest A&F University, Master
2015.09-2019.12, Northwest A&F University, Ph.D
Working experience:
2018.11-2020.10, Wageningen University & Research, Postdoc
2020.11 until now, Northwest A&F University, Associate professor
Wang Qian
E-mail: [email protected]
Tel: 15249210684
Research Interests:
Agricultural economic management, Land resource management
1
Course requirements
• Compulsory course
2
Questions
• What did you learn in the last six weeks?
• Profit maximization
• Cost minimization
• Consumer theory
3
Contents
• Public goods
• Externalities
• Behavioral economics
4
Chapter 15 Game Theory
Qian Wang
E-mail: [email protected]
Telephone: 15249210684
5
Contents
• 15.1 Introduction
• 15.2 Solution
• Solution concepts
• Nash equilibrium
7
15.1 Introduction
Questions
a) How much do you know about game theory?
Rock-paper-scissors
Tian Ji, a general of Qi State, had a penchant for racing
horses. Once he made a deal with Duke Wei of Qi
State to have a horse race. They categorized their
horses into three classes: top, middle and low. This
time, both of them sent in the top, middle and low
classes of horses in order. As the horses of Duke were
better than those of Tian in every class, Tian lost8 all
Horse racing
three races.
15.1 Introduction
More examples:
War
Marriage
Gambling
College admission
……
9
15.1 Introduction
• Game theory was originally developed by the
Hungarian-born American mathematician John von
Neumann and his Princeton University colleague
Oskar Morgenstern, a German-born American
economist, to solve problems in economics.
• Game theory was further developed in the 1950s by
American mathematician John Nash, who
established the mathematical principles of game
theory, a branch of mathematics that examines the
rivalries between competitors with mixed interests.
13
15.1 Introduction
• Game matrix
14
15.1 Introduction
Example 1:Matching Pennies
15
15.1 Introduction
Example 2:The Prisoners’ Dilemma
• There were two prisoners who jointly
participated in a crime. They could
cooperate with each other and refuse to
give evidence, or one could defect and
implicate the other.
16
15.1 Introduction
Example 3:Cournot Duopoly
• We suppose that there are two firms who produce an identical good at zero
cost. Each firm must decide how much output to produce without knowing
the production decision of the other duopolist.
• If the firms produce a total of x units of the good, the market price will be
p(x); that is, p(x) is the inverse demand curve facing these two producers.
• If 𝑥𝑖 is the production level of firm i, the market price will then be 𝑝(𝑥1 +
𝑥2 ), and the profits of firm i are given by 𝑛𝑖 = 𝑝 𝑥1 + 𝑥2 𝑥𝑖 . In this game the
strategy of firm i is its choice of production level and the payoff to firm i is
its profits.
17
15.1 Introduction
Example 4:Bertrand Duopoly
• Suppose that the strategy of each player is to announce the price at which
he would be willing to supply an arbitrary amount of the good in question.
• It is plausible to suppose that the consumers will only purchase from the
firm with the lowest price, and that they will split evenly between the two
firms if they charge the same price.
This game has a similar structure to that of the Prisoner's Dilemma. If both
players cooperate, they can charge the monopoly price and each reap half
of the monopoly profits. But the temptation is always there for one player to
cut its price slightly and thereby capture the entire market for itself. But if
both players cut price, then they are both worse off.
18
15.1 Introduction
Economic modeling of strategic choices
• In the Cournot game, the payoff to each firm is a continuous function of its
strategic choice;
• In the Bertrand game, the payoffs are discontinuous functions of the strategies.
19
• 15.2 Solution
20
15.2 Solution
Solution concepts
• Neither player wants the other player to be able to predict his choice
accurately.
21
15.2 Solution
• Exercise:
22
15.2 Solution
• Exercise:
23
15.2 Solution
Solution concepts
• Pure strategy
• If R is the set of pure strategies available to Row, the set of mixed
strategies open to Row will be the set of all probability distributions over R,
where the probability of playing strategy r in R is 𝑝𝑟 . Similarly, 𝑝𝑐 , will be
the probability that Column plays some strategy c.
• In order to solve the game, we want to find a set of mixed strategies (𝑝𝑟 , 𝑝𝑐 )
that are in equilibrium. It may be that some of the equilibrium mixed
strategies assign probability 1 to some choices, in which case they are
interpreted as pure strategies.
24
15.2 Solution
Solution concepts
• Each wants to maximize his or her expected utility given his or her beliefs.
Given my beliefs about what the other player might do, I choose the
strategy that maximizes my expected utility.
• Each player in the game knows that the other player is out to maximize his
own payoff, and each should use that information in determining what are
reasonable beliefs to have about the other player's behavior.
26
15.2 Solution
• One of the fundamental principles of game theory, the idea of
equilibrium strategies was developed by John F. Nash.
27
15.2 Solution
Nash equilibrium
• A Nash equilibrium consists of probability beliefs, (𝜋𝑟 , 𝜋𝑐 ) over strategies, and
probability of choosing strategies (𝑝𝑟 , 𝑝𝑐 ), such that:
• I) the beliefs are correct: 𝑝𝑟 = 𝜋𝑟 and 𝑝𝑐 = 𝜋𝑐 , for all r and c; and,
• 2) each player is choosing (𝑝𝑟 ) and (𝑝𝑐 ) so as to maximize his expected utility given
his beliefs.
• In equilibrium each player correctly foresees how likely the other player is to
make various choices, and the beliefs of the two players are mutually consistent.
• It is a pair of mixed strategies (𝑝𝑟 , 𝑝𝑐 ) such that each agent's choice maximizes
his expected utility, given the strategy of the other agent.
28
15.2 Solution
Nash equilibrium
• Pure strategy Nash equilibrium
29
15.2 Solution
Nash equilibrium
• If Row believes that Column will play c*, then Row’s best reply is r* and
similarly for Column. No player would find it in his or her interest to deviate
unilaterally from a Nash equilibrium strategy.
• If a set of strategies is not a Nash equilibrium then at least one player is not
consistently thinking through the behavior of the other player. That is, one of
the players must expect the other player not to act in his own self-interest——
ontradicting the original hypothesis of the analysis.
• Row might think: "If I think that Column is going to play some strategy 𝑐1 then
the best response for me is to play 𝑟1 . But if Column thinks that I will play 𝑟1 ,
then the best thing for him to do is to play some other strategy 𝑐2 . But if Column
is going to play 𝑐2 , then my best response is to play 𝑟2 . . ." and so on.
• A Nash equilibrium is then a set of beliefs and strategies in which each player's
beliefs about what the other player will do are consistent with the other player's
actual choice.
31
15.2 Solution
Nash equilibrium
• Example: Battle of the Sexes
32
15.2 Solution
Nash equilibrium
• Exercise1(5 minutes):
• Two California teenagers Bill and
Ted are playing Chicken. Bill
drives his hot rod south down a
one-lane road, and Ted drives his
hot rod north along the same Questions:
road. Each has two strategies: (a) Find all pure strategy equilibria.
Stay or Swerve. If one player
chooses Swerve he looses face; if
both Swerve, they both lose face.
However, if both choose Stay,
they are both killed. The payoff
matrix for Chicken looks like this:
33
15.2 Solution
Nash equilibrium
• Mixed strategy Nash equilibrium for battle of sexes
• Let (𝑝𝑡 , 𝑝𝑏 ) be the probabilities with which Row plays Top and Bottom, and
define (𝑝𝑙 , 𝑝𝑟 ) in a similar manner. Then Row's problem is
34
15.2 Solution
• Since we already know the pure strategy solutions, we only consider the
case where 𝑝𝑡 > 0 and 𝑝𝑏 > 0. The complementary slackness conditions
then imply that 𝑝𝑡 = 𝑝𝑏 = 0. Using the fact that 𝑝𝑙 + 𝑝𝑟 = 1, we easily see
that Row will find it optimal to play a mixed strategy when 𝑝𝑙 =
1Τ 𝑎𝑛𝑑 𝑝 = 2Τ .
3 𝑟 3
• Following the same procedure for Column, we find that 𝑝𝑡 = 2Τ3 and 𝑝𝑏 =
1Τ .
3
35
15.2 Solution
Nash equilibrium
• Exercise1(10 minutes)
• Two California teenagers Bill and
Ted are playing Chicken. Bill drives
his hot rod south down a one-lane
road, and Ted drives his hot rod Questions:
north along the same road. Each (b) Find all mixed strategy equilibria.
has two strategies: Stay or Swerve. (c) What is the probability that both teenagers
If one player chooses Swerve he will survive.
looses face; if both Swerve, they
both lose face. However, if both
choose Stay, they are both killed.
The payoff matrix for Chicken
looks like this:
36
15.2 Solution
• 1) Will a Nash equilibrium generally exist;
Nash (1950) showed that with a finite number of agents and a finite number
of pure strategies, an equilibrium will always exist.
37
15.2 Solution
• Dominant strategy
• Let 𝑟1 and 𝑟2 be two of Row‘s strategies. We say that 𝑟1 strictly dominates 𝑟2 for
Row if the payoff from strategy 𝑟1 is strictly larger than the payoff for 𝑟2 no matter
what choice Column makes.
• The strategy 𝑟1 weakly dominates 𝑟2 if the payoff from 𝑟1 is at least as large for
all choices Column might make and strictly larger for some choice.
• Note: A dominant strategy equilibrium is a Nash equilibrium, but not all Nash equilibria
are dominant strategy equilibria. 38
15.2 Solution
• Elimination of dominated strategies
• When there is no dominant strategy equilibrium, we have to resort to the
idea of a Nash equilibrium. But typically there will be more than one Nash
equilibrium. Our problem then is to try to eliminate some of the Nash
equilibria as being "unreasonable."
When given a game, we should first eliminate all strategies that are
dominated and then calculate the Nash equilibria of the remaining game.
This procedure is called elimination of dominated strategies; it can
sometimes result in a significant reduction in the number of Nash equilibria.
39
15.2 Solution
• Example: Elimination of dominated strategies
41
• 15.3 Repeated game
42
15.3 Repeated games
• Give me one more time, I will……
• Next time, I will……
• It was not appropriate to expect that the outcome of a repeated game with
the same players as simply being a repetition of the one-shot game.
• Each player can determine his or her choice at some point as a function of
the entire history of the game up until that point.
• The payoffs in the repeated game are the discounted sums of the payoffs at
each stage. If a player gets a payoff at time t of 𝑢𝑡 , his payoff in the repeated
game is taken to be σ∞ 𝑡
𝑡=0 𝑢𝑡 (1 + 𝑟) ,where r is the discount rate.
• As long as the discount rate is not too high there exists a Nash equilibrium pair
of strategies such that each player finds it in his interest to cooperate at each
stage.
44
15.3 Repeated games
Cooperate on the current move unless the other player defected on the last
move. If the other player defected on the last move, then Defect forever.
• Punishment strategy
• In a repeated Prisoner’s Dilemma any payoff larger than the payoff received if
both parties consistently defect can be supported as a Nash equilibrium.
• Maintaining a cartel
45
• 15.4 Sequential games
46
15.4 Sequential games
In many situations at least some of the choices are made
sequentially, and one player may know the other player's choice
before he has to make his own choice.
47
15.4 Sequential games
Game tree: This is simply a diagram that indicates the choices that each
player can make at each point in time. The payoffs to each player are
indicated at the "leaves" of the tree.
48
15.4 Sequential games
Once a choice has been made, the players are in a subgame consisting of the
strategies and payoffs available to them from then on.
Only one of the two Nash equilibria satisfies the condition that it is not only an
overall equilibrium, but also an equilibrium in each of the subgames. A Nash
equilibrium with this property is known as a subgame perfect equilibrium.
49
15.4 Sequential games
The extensive form of the game is also capable of modeling situations where
some of the moves are sequential and some are simultaneous.
54
15.5 Repeated games and subgame perfection
• A somewhat less harsh strategy is known as Tit-for-Tat: start out
cooperating on the first play and on subsequent plays do whatever your
opponent did on the previous play. In this strategy, a player is punished for
defection, but he is only punished once. In this sense Tit-for-Tat is a
"forgiving“ strategy.
55
• 15.6 Games with incomplete information
56
15.6 Games with incomplete information
• If one agent doesn't know the payoffs of the other agent, then the Nash
equilibrium doesn't make much sense.
• One approach is to subsume all of the uncertainty that one agent may have about
another into a variable known as the agent’s type.
• For example, one agent may be uncertain about another agent’s valuation of some good,
about his or her risk aversion and so on. Each type of player is regarded as a different player
and each agent has some prior probability distribution defined over the different types of
agents.
58