1 / 36

Unit III: The Evolution of Cooperation

Unit III: The Evolution of Cooperation. Can Selfishness Save the Environment? Repeated Games: the Folk Theorem Evolutionary Games A Tournament How to Promote Cooperation. 7/28. 4/14. 7/31. How to Promote Cooperation. Advice to Participants/Reformers The Logic of Collective Action

gamma
Download Presentation

Unit III: The Evolution of Cooperation

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Unit III: The Evolution of Cooperation • Can Selfishness Save the Environment? • Repeated Games: the Folk Theorem • Evolutionary Games • A Tournament • How to Promote Cooperation 7/28 4/14 7/31

  2. How to Promote Cooperation • Advice to Participants/Reformers • The Logic of Collective Action • Changing the Rules of the Game • The Problem of Trust • Limits of Strategic Rationality • Tournament Update

  3. How to Promote Cooperation Based on the success of TFT in his tournaments, Axelrod offers two sets of recommendations to promote cooperation (1984, pp. 199-244): Advice to Participants - How to do well as a player in the game; the ingredients of a strategy that will do well in IRPD. Advice to Reformers - How to change the rules of the game to promote cooperation; changing the rules changes the players payoffs and hence the game.

  4. Advice to Participants How to Choose Effectively (Axelrod, 1984, pp. 109-123.) • Don’t be envious • Don’t be the first to defect • Reciprocate both cooperation and defection • Don’t be too clever These are intended as the ingredients of a strategy that will, in the long range and against a wide range of opponents, advance the player’s interests.

  5. Advice to Participants • Nice: Never be the first to defect. A nice strategy signals a willingness to cooperate and may induce reciprocal cooperation. Nice strategies did best in Axelrod’s tournaments. • Provocable: Punish defection. Don’t get fleeced by an exploiter. • Forgiving: Reciprocate cooperation. Triggers may be susceptible to misunderstandings, mistakes, etc, that can lead otherwise cooperative players into spirals of alternating or mutual defection. • Clear: Be easily understood and predictable. Don’t be too clever. A simple rule works best.

  6. Advice to Participants Sucker the Simple? Recall that while TIT FOR TAT never beats is opponent, PAVLOV always defects against a naïve cooperator. Hence, the success of PAVLOV in newer tournaments may suggest it is wise to exploit the weak, both (i) for “egoistic” benefit; and (ii) to increase the overall fitness of the population. Either the simple will learn (not to let themselves be exploited), or they will be winnowed.

  7. Advice to Reformers Axelrod offers five concrete suggestions on how “the strategic setting itself can be transformed in order to promote cooperation among the players” (1984, pp. 124-141): • Enlarge the “shadow of the future” • Change the payoffs • Teach people to care about each other • Teach reciprocity • Improve recognition abilities

  8. Advice to Reformers Repeated interactions provide the conditions necessary for cooperation by transforming the nature of the interaction in two ways: • “Enlarge the shadow of the future” • Increase the amount of information in the system. This may reduces strategic uncertainty (e) and allow players to coordinate their expectations and behavior on mutually beneficial outcomes. d 1 d* = T-R T-P e0

  9. Unit IV: Thinking about Thinking • Behavioral Game Theory • Learning to Cooperate • Summary and Conclusions

  10. Behavioral Game Theory • The Limits of Homo Economicus • Bounded Rationality • Learning to Cooperate • Tournament Update

  11. The Limits of Homo Economicus Schelling’s “Errant Economics” “The Intimate Contest for Self-Command” (1984: 57-82) “The Mind as a Consuming Organ” (328-46) The standard model of rational economic man is: • Too simple • Assumes time consistent preferences • Susceptible to self deception and ‘sour grapes’ • Is overly consequentialist • Ignores ‘labelling’ and ‘framing’ effects

  12. The Limits of Homo Economicus Schelling’s “Errant Economics” “The Intimate Contest for Self-Command” (1984: 57-82) “The Mind as a Consuming Organ” (328-46) Schelling’s views are not merely critical (negative); his concerns foreshadow much current research on improving the standard model: • Behavioral economics/cognitive psychology • Artificial Intelligence • Learning models: Inductive reasoning

  13. The Limits of Homo Economicus Experiments in “behavioral economics” have shown people routinely do not behave the way the standard model predicts: • reject profitable bargains they think are unfair • do not take full advantage of others when they can • punish others even when costly to themselves • contribute substantially to public goods • behave irrationally when they expect others to behave even more irrationally (Camerer, 1997)

  14. Bounded Rationality Game theory usually assumes “unbounded,” perfect, or “Olympian” rationality (Simon, 1983). Players: • have unlimited memory and computational resources. • solve complex, interdependent maximization problems – instantaneously! – subject only to the constraint that the other player is also trying to maximize. But observation and experimentation with human subjects tell us that people don’t actually make decisions this way. A more realistic approach would make more modest assumptions: bounded rationality.

  15. Bounded Rationality Game theory usually assumes players are deductively rational. Starting from certain givens (sets of actions, information, payoffs), they arrive at a choice that maximizes expected utility. Deductiverationality assumes a high degree of constancy in the decision-makers’ environment. They may have complete or incomplete information, but they are able to form probability distributions over all possible states of the world, and these underlying distributions are themselves stable. But in more complex environments, the traditional assumptions break down. Every time a decision is made the environment changes, sometimes in unpredictable ways, and every new decision is made in a new environment (S. Smale).

  16. Bounded Rationality In more complicated environments, the computational requirements to deduce a solution quickly swamp the capacity of any human reasoning. Chess appears to be well beyond the ability of humans to fulfill the requirements of traditional deductive reasoning. In today’s “fast” economy a more dynamic theory is needed. The long-run position of the economy may be affected by our predictions! “On Learning and Adaptation in the Economy,” Arthur, 1992, p. 5

  17. Bounded Rationality • The standard model of Homo Economics break down for two reasons: • (i) human decision making is limited by finite memory and computational resources. • (ii) thinking about others’ thinking involves forming subjective beliefs and subjective beliefs about subjective beliefs, and so on.

  18. Bounded Rationality There is a peculiar form of regress which characterizes reasoning about someone else’s reasoning, which in turn, is based on assumptions about one's own reasoning, a point repeatedly stressed by Schelling (1960). In some types of games this process comes to an end in a finite number of steps . . . . Reflexive reasoning, . . . ‘folds in on itself,’ as it were, and so is not a finite process. In particular when one makes an assumption in the process of reasoning about strategies, one ‘plugs in’ this very assumption into the ‘data.’ In this way the possibilities may never be exhausted in a sequential examination. Under these circumstances it is not surprising that the purely deductive mode of reasoning becomes inadequate when the reasoners themselves are the objects of reasoning. (Rapoport, 1966, p. 143)

  19. Bounded Rationality In the Repeated Prisoner’s Dilemma, it has been suggested that “uncooperative behavior is the result of ‘unbounded rationality’, i.e., the assumed availability of unlimited reasoning and computational resources to the players” (Papadimitrou, 1992: 122). If players are bounded rational, on the other hand, the cooperative outcome may emerge as the result of a “muddling” process. They reason inductively and adapt (imitate or learn) locally superior stategies. Thus, not only is bounded rationality a more “realistic” approach, it may also solve some deep analytical problems, e.g., resolution of finite horizon paradoxes.

  20. Learning to Cooperate The Folk Theorem The shaded area is the set of SPNE. The segment PP,RR is the set of “collectively stable” strategies, for (d > d*). (S,T) (R,R) (P,P) (T,S)

  21. Learning to Cooperate We have seen that whereas cooperation is irrational in a one-shot Prisoner’s Dilemma, it may be rational (i.e., achieved in a SPNE), if the game is repeated and “the shadow of the future” is sufficiently large: d > (T-R)/(T-P) (i) Repeated interaction is a necessary but not a sufficient condition for cooperation. In addition, players must have reason to believe the other will reciprocate. This involves judging intentions, considerations of fairness, (mis)communication, trust, deception, etc.

  22. Learning to Cooperate Consider two fishermen deciding how many fish to remove from a commonly owned pond. There are Y fish in the pond. • Period 1 each fishery chooses to consume (c1, c2). • Period 2 remaining fish are equally divided (Y – (c1+c2))/2). c1 = (Y – c2)/2 NE: c1 = c2 = Y/3 Social Optimum: c1 = c2 = Y/4 c2 Y/3 Y/4 c2 = (Y – c1)/2 Y/4Y/3 c1

  23. Learning to Cooperate Consider two fishermen deciding how many fish to remove from a commonly owned pond. There are Y fish in the pond. • Period 1 each fishery chooses to consume (c1, c2). • Period 2 remaining fish are equally divided (Y – (c1+c2))/2). c1 = (Y – c2)/2 If there are 12 fish in the pond, each will consume (Y/3) 4 in the spring and 2 in the fall in a NE. Both would be better off consuming (Y/4) 3 in the fall, leaving 3 for each in the spring. c2 Y/3 Y/4 c2 = (Y – c1)/2 Y/4Y/3 c1

  24. Learning to Cooperate If there are 12 fish in the pond, each will consume (Y/3) 4 in the spring and 2 in the fall in a NE. Both would be better off consuming (Y/4) 3 in the fall, leaving 3 for each in the spring. C D C9, 9 7.5,10 A Prisoner’s Dilemma What would happen if the game were repeated? D 10,7.5 8, 8

  25. Learning to Cooperate Imagine the fisherman make the following deal: Each will Cooperate (consume only 3) in the spring as long as the other does likewise; as soon as one Defects, the other will Defect for ever, i.e., they adopt trigger strategies. This deal will be stable if the threat of future punishment makes both unwilling to Defect, i.e., if the one period gain from Defect is not greater than the discounted future loss due to the punishment: (T – R) < (dR/(1-d) – dP/(1-d)) (ii)

  26. Learning to Cooperate Imagine the fishermen make the following deal: Promise: I’ll consume 3 in the spring, if you do. Threat: I’ll consume 4, forever, if you deviate. 9 … 9 … 9 … 9 … 9 … 9 … 9 … 9 … = 9/(1-d) 9 … 9 … 9 … 9 … 10 … 8 … 8 … 8 … If d is sufficiently high, the threat will be credible, and the pair of trigger strategies is a Nash equilibrium. d* = 0.5 Trigger Strategy Current gain from deviation = 10 – 9 = 1 Future gain from cooperation = d9/(1-d) –d8/(1-d)

  27. Learning to Cooperate Imagine the fishermen make the following deal: Promise: I’ll consume 3 in the spring, if you do. Threat: I’ll consume 4, forever, if you deviate. R … R … R … R … R … R … R … R … = R/(1-d) R … R … R … R … T … P … P … P … If d is sufficiently high, the threat will be credible, and the pair of trigger strategies is a Nash equilibrium. d* = (T-R)/(T-P) Trigger Strategy Current gain from deviation = (T – R) Future gain from cooperation = dR/(1-d)-dP/(1-d)

  28. Learning to Cooperate Imagine the fisherman make the following deal: Each will Cooperate (consume only 3) in the spring as long as the other does likewise; as soon as one Defects, the other will Defect for ever, i.e., they adopt trigger strategies. This deal will be stable if the threat of future punishment makes both unwilling to Defect, i.e., if the one period gain from Defect is not greater than the discounted future loss due to the punishment: (T – R) < (dR/(1-d) – dP/(1-d)) or d> (T-R)/(T-P)

  29. Learning to Cooperate Imagine there are many fishermen, each of whom can adopt either D(efect), C(ooperate), or T(rigger). In every generation, each fisherman plays against every other. After each generation, those that did poorly can switch to imitate those that did better. Eventually, C will die out, and the population will be dominated by either D or T, depending on the discount parameter. Noise (miscommunication) can also affect the outcome.

  30. Tournament Assignment Design a strategy to play an Evolutionary Prisoner’s Dilemma Tournament. Entries will meet in a round robin tournament, with 1% noise (i.e., for each intended choice there is a 1% chance that the opposite choice will be implemented). Games will last at least 1000 repetitions (each generation), and after each generation, population shares will be adjusted according to the replicator dynamic, so that strategies that do better than average will grow as a share of the population whereas others will be driven to extinction. The winner or winners will be those strategies that survive after at least 10,000 generations.

  31. Tournament Assignment To design your strategy, access the programs through your fas Unix account. The Finite Automaton Creation Tool (fa) will prompt you to create a finite automata to implement your strategy. Select the number of internal states, designate the initial state, define output and transition functions, which together determine how an automaton “behaves.” The program also allows you to specify probabilistic output and transition functions. Simple probabilistic strategies such as GENEROUS TIT FOR TAT have been shown to perform particularly well in noisy environments, because they avoid costly sequences of alternating defections that undermine sustained cooperation.

  32. Preliminary Tournament Results After 5000 generations (as of 4/25/02) Avg. Score (x10)

  33. Preliminary Tournament Results After 5000 generations (10pm 4/27/02)

  34. Preliminary Tournament Results After 20000 generations (7am 4/28/02)

  35. Preliminary Tournament Results After 600 generations (4/22/05)

  36. Preliminary Tournament Results After 1000 generations (4/27/05)

More Related