The “classic” models of oligopoly (Cournot & Bertrand) have significant flaws
Cournot ignored the possibility of collusion (we considered it alongside Cournot's model)
Consider a profit possibilities frontier between Coke and Pepsi should they collude
Point C: Cournot-Nash equilibrium
Points Mc, Mp, if Coke or Pepsi were a monopolist, respectively
Anything northeast of C is a Pareto improvement (for the firms)
A bargaining problem between Coke and Pepsi
Coke would prefer point E, Pepsi point D, point F is a 50:50 split
But in any case, a lot of room for a mutually-beneficial agreement to cooperate instead of (Cournot) competing
George Stigler
1911—1991
Economics Nobel 1982
“The present paper accepts the hypothesis that oligopolists wish to collude to maximize joint profits. It seeks to reconcile this wish with the facts, such as that collusion is impossible for many firms and collusion is much more effective in some circumstances than in others. The reconciliation is found in the problem of policing a collusive agreement, which proves to be a problem in the theory of information,” (44).
Stigler, George J, 1964, “A Theory of Oligopoly,” Journal of Political Economy 72(1): 44-61
George Stigler
1911—1991
Economics Nobel 1982
“We shall show that collusion normally involves much more than ‘the’ price...The colluding firms must agree upon the price structure appropriate to the transaction classes which they are prepared to recognize. A complete profit-maximizing price structure may have almost infinitely numerous price classes: the firms will have to decide upon the number of price classes in the light of the costs and returns from tailoring prices to the diversity of transactions,” (44-46).
George Stigler
1911—1991
Economics Nobel 1982
“Let us assume that the collusion has been effected, and a price structure agreed upon. It is a well-established proposition that if any member of the agreement can secretly violate it, he will gain larger profits than by conforming to it. It is, moreover, surely one of the axioms of human behavior that all agreements whose violation would be profitable to the violator must be enforced. The literature of collusive agreements...is replete with instances of the collapse of conspiracies because of ‘secret’ price-cutting. This literature is biased: conspiracies that are successful in avoiding an amount of price-cutting which leads to collapse of the agreement are less likely to be reported or detected. But no conspiracy can neglect the problem of enforcement,” (46)
George Stigler
1911—1991
Economics Nobel 1982
“Enforcement consists basically of detecting significant deviations from the agreed-upon prices. Once detected, the deviations will tend to disappear because they are no longer secret and will be matched by fellow conspirators if they are not withdrawn. If the enforcement is weak, however — if price-cutting is detected only slowly and incompetently — the conspiracy must recognize its weakness: it must set prices not much above the competitive level so the inducements to price-cutting are small...” (46).
George Stigler
1911—1991
Economics Nobel 1982
“Policing the collusion sounds very much like the subtle and complex problem presented in a good detective story. [But] there is a difference: In our case the man who murders the collusive price will recieve the bequest of patronage. The basic method of detection of a price-cutter must be the fact that he is getting business he would not otherwise obtain,” (47).
George Stigler
1911—1991
Economics Nobel 1982
We should focus oligopoly theory less on static models of Cournot/Bertrand/etc competition
Focus more on examining the types of conditions where firms can effectively form and maintain collusive agreements, and conditions where agreements break down into competition
Consider more of a dynamic game of cooperation and/or competition between firms
See my game theory course for more
There's a lot more to game theory than a one-shot prisoners' dilemma:
one shot vs. repeated game
discrete vs. continuous strategies
perfect vs. imperfect vs. incomplete/asymmetric information
simultaneous vs. sequential games
We use various “solution concepts” to allow us to predict an equilibrium of a game
Nash Equilibrium is the primary solution concept
Recall, Nash Equilibrium: no players want to change their strategy given what everyone else is playing
N.E. ≠ the “best” or optimal outcome
Game may have multiple N.E.
Game may have no N.E. (in “pure” strategies)
1) Cell-by-Cell Inspection: look in each cell, does either player want to deviate?
2) Best-Response Analysis: take the perspective of each player. If the other player plays a particular strategy, what is your strategy(s) that gets you the highest payoff?
Player 1's best responses
2) Best-Response Analysis: take the perspective of each player. If the other player plays a particular strategy, what is your strategy(s) that gets you the highest payoff?
Player 2's best responses
2) Best-Response Analysis: take the perspective of each player. If the other player plays a particular strategy, what is your strategy(s) that gets you the highest payoff?
N.E.: each player is playing a best response
2) Best-Response Analysis: take the perspective of each player. If the other player plays a particular strategy, what is your strategy(s) that gets you the highest payoff?
Two Nash equilibria again: (A,A) and (B,B)
But here (A,A) ≻ (B,B)!
Path Dependence: early choices may affect later ability to choose or switch
Lock-in: the switching cost of moving from one equilibrium to another becomes prohibitive
Suppose we are currently in equilibrium (B,B)
Inefficient lock-in:
Consider a sequential game of Cournot competition between Coke and Pepsi
Coke moves first, then Pepsi, then the game ends
Each player can:
Designing a game tree:
Decision nodes: decision point for each player
Terminal nodes: outcome of game, with payoff for each player (profits)
(“Pure”) strategy†: a player’s complete plan of action for every possible contingency
Think of a strategy like an algorithm:
If we reach node 1, then I will play X; if we reach node 2, then I will play Y; if...
† “Pure” is meant to contrast against “mixed” strategies, where players take a range of actions according to a probability distribution. That's beyond the scope of this class.
Coke has 21=2 possible strategies:
Pepsi has 22=4 possible strategies:
Solve a sequential game by “backward induction” or “rollback”
To determine the outcome of the game, start with the last-mover (i.e. decision nodes just before terminal nodes) and work to the beginning
A process of considering “sequential rationality”:
“If I play X, my opponent will respond with Y; given their response, do I really want to play X?”
We start at P.1 where Pepsi can:
And P.2 where Pepsi can:
Pepsi will Defect if the game reaches node P.1 and Defect if the game reaches node P.2
Recognizing this, what will Coke do?
(Defect, (Defect, Defect))
As we work backwards, we can prune the branches of the game tree
Equilibrium path of play is highlighted from the root to one terminal node
A true prisoners' dilemma: a>b>c>d
Each player's preferences:
Nash equilibrium: (Defect, Defect)
We'll stick with these specific payoffs for this lesson
How can we sustain cooperation in Prisoners' Dilemma?
Analysis of games can change when players encounter each other more than once
Repeated games: the same players play the same game multiple times, two types:
Players know the history of the game with each other
Finitely-repeated game: has a known final round
Infinitely-repeated game: has no (or an unknown) final round
Suppose a prisoners' dilemma is played for 2 rounds
Apply backwards induction:
Suppose a prisoners' dilemma is played for 2 rounds
Apply backwards induction:
Suppose a prisoners' dilemma is played for 2 rounds
Apply backwards induction:
Suppose a prisoners' dilemma is played for 2 rounds
Apply backwards induction:
Both Defect in round 1 (and round 2)
No value in cooperation over time!
Paradox of repeated games: for any game with a unique Nash equilibrium (in pure strategies) in a one-shot game, as long as there is a known, finite end, Nash equilibrium is the same
Sometimes called Selten’s “chain-store paradox” from a famous paper by Reinhard Selten (1978)
In experimental settings, we tend to see people cooperate in early rounds, but close to the final round (if not the actual final round), defect on each other
Selten, Reinhard, (1978), “The chain store paradox,” Theory and Decision 9(2): 127–159
Finitely-repeated games are interesting, but rare
Some predictions for finitely-repeated games don't hold up well in reality
We often play games or are in relationships that are indefinitely repeated (have no known end), we call them infinitely-repeated games
Since we are dealing with payoffs in the future, we have to consider players' time preferences
Easiest to consider with monetary payoffs and the time value of money that underlies finance
PV=FV(1+r)t
FV=PV(1+r)t
PV=FV(1+r)nPV=1000(1+0.05)1PV=10001.05PV=$952.38
FV=PV(1+r)nFV=1000(1+0.05)1FV=1000(1.05)FV=$1,050
Suppose a player values $1 now as being equivalent to some amount with interest 1(1+r) one period later
The “discount factor” is δ=11+r, the ratio that future value must be multiplied to equal present value
$1 now=δ$1 later
If δ is low (r is high)
If δ is high (r is low)
Example: Suppose you are indifferent between having $1 today and $1.10 next period
Example: Suppose you are indifferent between having $1 today and $1.10 next period
$1 today=δ$1.10 next period$1$1.10=δ0.91≈δ
Example: Suppose you are indifferent between having $1 today and $1.10 next period
$1 today=δ$1.10 next period$1$1.10=δ0.91≈δ
There is an implied interest rate of r=0.10
$1 at 10% interest yields $1.10 next period
δ=11+rδ=11.10 ≈0.91
p(δ1+δ2+δ3+⋯+δT)
p(δ1+δ2+δ3+⋯+δT)
∞∑t=1=p1−δ
Alternate interpretation: game continues with some (commonly known among the players) probability θ each round
Assume this probability is independent between rounds (i.e. one round continuing has no influence on the probability of the next round continuing, etc)
Then the probability the game is played T rounds from now is θT
A payoff of p in every future round has a present value of p(θ1+θ2+θ3+⋯)=(p1−θ)
Note this is similar to discounting of future payoffs (at a constant rate); equivalent if θ=δ
Recall, a strategy is a complete plan of action that describes how you will react under all possible circumstances (i.e. moves by other players)
For an infinitely-repeated game, an infinite number of possible strategies exist!
We will examine a specific set of contingent or trigger strategies
Consider one (the most important) trigger strategy for an infinitely-repeated prisoners' dilemma, the “Grim Trigger” strategy:
“Grim” trigger strategy leaves no room for forgiveness: one deviation triggers infinite punishment, like the sword of Damocles
50+50δ1+50δ2+50δ3+⋯+50δ∞=501−δ
Payoff to cooperation>Payoff to one-time defection501−δ>57+45δ1−δδ>0.583
In general, can sustain cooperation (under grim trigger strategy) when δ>b−ac−a
Thus, cooperation breaks down (or is strengthened) when:
Robert Axelrod
1943—
“The contestants ranged from a 10-year-old computer hobbyist to professors of computer science, economics, psychology, mathematics, sociology, political science, and evolutionary biology.”
1 Each round had a 0.00346 probability of ending the game, ensuring on average 200 rounds of play
Axelrod, Robert, 1984, *The Evolution of Cooperation
Robert Axelrod
1943—
Axelrod's discussion of successful strategies based on four properties:
The winning strategy was, famously, TIT FOR TAT, submitted by Anatol Rapoport
Keyboard shortcuts
↑, ←, Pg Up, k | Go to previous slide |
↓, →, Pg Dn, Space, j | Go to next slide |
Home | Go to first slide |
End | Go to last slide |
Number + Return | Go to specific slide |
b / m / f | Toggle blackout / mirrored / fullscreen mode |
c | Clone slideshow |
p | Toggle presenter mode |
t | Restart the presentation timer |
?, h | Toggle this help |
o | Tile View: Overview of Slides |
Esc | Back to slideshow |
The “classic” models of oligopoly (Cournot & Bertrand) have significant flaws
Cournot ignored the possibility of collusion (we considered it alongside Cournot's model)
Consider a profit possibilities frontier between Coke and Pepsi should they collude
Point C: Cournot-Nash equilibrium
Points Mc, Mp, if Coke or Pepsi were a monopolist, respectively
Anything northeast of C is a Pareto improvement (for the firms)
A bargaining problem between Coke and Pepsi
Coke would prefer point E, Pepsi point D, point F is a 50:50 split
But in any case, a lot of room for a mutually-beneficial agreement to cooperate instead of (Cournot) competing
George Stigler
1911—1991
Economics Nobel 1982
“The present paper accepts the hypothesis that oligopolists wish to collude to maximize joint profits. It seeks to reconcile this wish with the facts, such as that collusion is impossible for many firms and collusion is much more effective in some circumstances than in others. The reconciliation is found in the problem of policing a collusive agreement, which proves to be a problem in the theory of information,” (44).
Stigler, George J, 1964, “A Theory of Oligopoly,” Journal of Political Economy 72(1): 44-61
George Stigler
1911—1991
Economics Nobel 1982
“We shall show that collusion normally involves much more than ‘the’ price...The colluding firms must agree upon the price structure appropriate to the transaction classes which they are prepared to recognize. A complete profit-maximizing price structure may have almost infinitely numerous price classes: the firms will have to decide upon the number of price classes in the light of the costs and returns from tailoring prices to the diversity of transactions,” (44-46).
George Stigler
1911—1991
Economics Nobel 1982
“Let us assume that the collusion has been effected, and a price structure agreed upon. It is a well-established proposition that if any member of the agreement can secretly violate it, he will gain larger profits than by conforming to it. It is, moreover, surely one of the axioms of human behavior that all agreements whose violation would be profitable to the violator must be enforced. The literature of collusive agreements...is replete with instances of the collapse of conspiracies because of ‘secret’ price-cutting. This literature is biased: conspiracies that are successful in avoiding an amount of price-cutting which leads to collapse of the agreement are less likely to be reported or detected. But no conspiracy can neglect the problem of enforcement,” (46)
George Stigler
1911—1991
Economics Nobel 1982
“Enforcement consists basically of detecting significant deviations from the agreed-upon prices. Once detected, the deviations will tend to disappear because they are no longer secret and will be matched by fellow conspirators if they are not withdrawn. If the enforcement is weak, however — if price-cutting is detected only slowly and incompetently — the conspiracy must recognize its weakness: it must set prices not much above the competitive level so the inducements to price-cutting are small...” (46).
George Stigler
1911—1991
Economics Nobel 1982
“Policing the collusion sounds very much like the subtle and complex problem presented in a good detective story. [But] there is a difference: In our case the man who murders the collusive price will recieve the bequest of patronage. The basic method of detection of a price-cutter must be the fact that he is getting business he would not otherwise obtain,” (47).
George Stigler
1911—1991
Economics Nobel 1982
We should focus oligopoly theory less on static models of Cournot/Bertrand/etc competition
Focus more on examining the types of conditions where firms can effectively form and maintain collusive agreements, and conditions where agreements break down into competition
Consider more of a dynamic game of cooperation and/or competition between firms
See my game theory course for more
There's a lot more to game theory than a one-shot prisoners' dilemma:
one shot vs. repeated game
discrete vs. continuous strategies
perfect vs. imperfect vs. incomplete/asymmetric information
simultaneous vs. sequential games
We use various “solution concepts” to allow us to predict an equilibrium of a game
Nash Equilibrium is the primary solution concept
Recall, Nash Equilibrium: no players want to change their strategy given what everyone else is playing
N.E. ≠ the “best” or optimal outcome
Game may have multiple N.E.
Game may have no N.E. (in “pure” strategies)
1) Cell-by-Cell Inspection: look in each cell, does either player want to deviate?
2) Best-Response Analysis: take the perspective of each player. If the other player plays a particular strategy, what is your strategy(s) that gets you the highest payoff?
Player 1's best responses
2) Best-Response Analysis: take the perspective of each player. If the other player plays a particular strategy, what is your strategy(s) that gets you the highest payoff?
Player 2's best responses
2) Best-Response Analysis: take the perspective of each player. If the other player plays a particular strategy, what is your strategy(s) that gets you the highest payoff?
N.E.: each player is playing a best response
2) Best-Response Analysis: take the perspective of each player. If the other player plays a particular strategy, what is your strategy(s) that gets you the highest payoff?
Two Nash equilibria again: (A,A) and (B,B)
But here (A,A) ≻ (B,B)!
Path Dependence: early choices may affect later ability to choose or switch
Lock-in: the switching cost of moving from one equilibrium to another becomes prohibitive
Suppose we are currently in equilibrium (B,B)
Inefficient lock-in:
Consider a sequential game of Cournot competition between Coke and Pepsi
Coke moves first, then Pepsi, then the game ends
Each player can:
Designing a game tree:
Decision nodes: decision point for each player
Terminal nodes: outcome of game, with payoff for each player (profits)
(“Pure”) strategy†: a player’s complete plan of action for every possible contingency
Think of a strategy like an algorithm:
If we reach node 1, then I will play X; if we reach node 2, then I will play Y; if...
† “Pure” is meant to contrast against “mixed” strategies, where players take a range of actions according to a probability distribution. That's beyond the scope of this class.
Coke has 21=2 possible strategies:
Pepsi has 22=4 possible strategies:
Solve a sequential game by “backward induction” or “rollback”
To determine the outcome of the game, start with the last-mover (i.e. decision nodes just before terminal nodes) and work to the beginning
A process of considering “sequential rationality”:
“If I play X, my opponent will respond with Y; given their response, do I really want to play X?”
We start at P.1 where Pepsi can:
And P.2 where Pepsi can:
Pepsi will Defect if the game reaches node P.1 and Defect if the game reaches node P.2
Recognizing this, what will Coke do?
(Defect, (Defect, Defect))
As we work backwards, we can prune the branches of the game tree
Equilibrium path of play is highlighted from the root to one terminal node
A true prisoners' dilemma: a>b>c>d
Each player's preferences:
Nash equilibrium: (Defect, Defect)
We'll stick with these specific payoffs for this lesson
How can we sustain cooperation in Prisoners' Dilemma?
Analysis of games can change when players encounter each other more than once
Repeated games: the same players play the same game multiple times, two types:
Players know the history of the game with each other
Finitely-repeated game: has a known final round
Infinitely-repeated game: has no (or an unknown) final round
Suppose a prisoners' dilemma is played for 2 rounds
Apply backwards induction:
Suppose a prisoners' dilemma is played for 2 rounds
Apply backwards induction:
Suppose a prisoners' dilemma is played for 2 rounds
Apply backwards induction:
Suppose a prisoners' dilemma is played for 2 rounds
Apply backwards induction:
Both Defect in round 1 (and round 2)
No value in cooperation over time!
Paradox of repeated games: for any game with a unique Nash equilibrium (in pure strategies) in a one-shot game, as long as there is a known, finite end, Nash equilibrium is the same
Sometimes called Selten’s “chain-store paradox” from a famous paper by Reinhard Selten (1978)
In experimental settings, we tend to see people cooperate in early rounds, but close to the final round (if not the actual final round), defect on each other
Selten, Reinhard, (1978), “The chain store paradox,” Theory and Decision 9(2): 127–159
Finitely-repeated games are interesting, but rare
Some predictions for finitely-repeated games don't hold up well in reality
We often play games or are in relationships that are indefinitely repeated (have no known end), we call them infinitely-repeated games
Since we are dealing with payoffs in the future, we have to consider players' time preferences
Easiest to consider with monetary payoffs and the time value of money that underlies finance
PV=FV(1+r)t
FV=PV(1+r)t
PV=FV(1+r)nPV=1000(1+0.05)1PV=10001.05PV=$952.38
FV=PV(1+r)nFV=1000(1+0.05)1FV=1000(1.05)FV=$1,050
Suppose a player values $1 now as being equivalent to some amount with interest 1(1+r) one period later
The “discount factor” is δ=11+r, the ratio that future value must be multiplied to equal present value
$1 now=δ$1 later
If δ is low (r is high)
If δ is high (r is low)
Example: Suppose you are indifferent between having $1 today and $1.10 next period
Example: Suppose you are indifferent between having $1 today and $1.10 next period
$1 today=δ$1.10 next period$1$1.10=δ0.91≈δ
Example: Suppose you are indifferent between having $1 today and $1.10 next period
$1 today=δ$1.10 next period$1$1.10=δ0.91≈δ
There is an implied interest rate of r=0.10
$1 at 10% interest yields $1.10 next period
δ=11+rδ=11.10 ≈0.91
p(δ1+δ2+δ3+⋯+δT)
p(δ1+δ2+δ3+⋯+δT)
∞∑t=1=p1−δ
Alternate interpretation: game continues with some (commonly known among the players) probability θ each round
Assume this probability is independent between rounds (i.e. one round continuing has no influence on the probability of the next round continuing, etc)
Then the probability the game is played T rounds from now is θT
A payoff of p in every future round has a present value of p(θ1+θ2+θ3+⋯)=(p1−θ)
Note this is similar to discounting of future payoffs (at a constant rate); equivalent if θ=δ
Recall, a strategy is a complete plan of action that describes how you will react under all possible circumstances (i.e. moves by other players)
For an infinitely-repeated game, an infinite number of possible strategies exist!
We will examine a specific set of contingent or trigger strategies
Consider one (the most important) trigger strategy for an infinitely-repeated prisoners' dilemma, the “Grim Trigger” strategy:
“Grim” trigger strategy leaves no room for forgiveness: one deviation triggers infinite punishment, like the sword of Damocles
50+50δ1+50δ2+50δ3+⋯+50δ∞=501−δ
Payoff to cooperation>Payoff to one-time defection501−δ>57+45δ1−δδ>0.583
In general, can sustain cooperation (under grim trigger strategy) when δ>b−ac−a
Thus, cooperation breaks down (or is strengthened) when:
Robert Axelrod
1943—
“The contestants ranged from a 10-year-old computer hobbyist to professors of computer science, economics, psychology, mathematics, sociology, political science, and evolutionary biology.”
1 Each round had a 0.00346 probability of ending the game, ensuring on average 200 rounds of play
Axelrod, Robert, 1984, *The Evolution of Cooperation
Robert Axelrod
1943—
Axelrod's discussion of successful strategies based on four properties:
The winning strategy was, famously, TIT FOR TAT, submitted by Anatol Rapoport