1 / 43

Reinforcement Learning Dynamics in Social Dilemmas

Reinforcement Learning Dynamics in Social Dilemmas. Luis R. Izquierdo & Segis Izquierdo. Outline of the presentation. BM reinforcement model Macy and Flache ’ s SRE and SCE Self-Reinforcing Equilibrium (SRE): Challenges Self-Correcting Equilibrium (SCE): Challenges. Motivation.

denzer
Download Presentation

Reinforcement Learning Dynamics in Social Dilemmas

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Reinforcement Learning Dynamics in Social Dilemmas Luis R. Izquierdo & Segis Izquierdo

  2. Outline of the presentation • BM reinforcement model • Macy and Flache’s SRE and SCE • Self-Reinforcing Equilibrium (SRE): Challenges • Self-Correcting Equilibrium (SCE): Challenges • Motivation • In-depth analysis of the dynamics of the model • Analysis of the robustness of the model • Conclusions

  3. BM reinforcement model • Reinforcement learners tend to repeat actions that led to satisfactory outcomes in the past, and avoid choices that resulted in unsatisfactory experiences. • The propensity or probability to play an action is increased (decreased) if it leads to a satisfactory (unsatisfactory) outcome.

  4. BM reinforcement model The Prisoner’s Dilemma pC= Probability to cooperate pD= Probability to defect Aspiration Threshold: A = 2 Learning rate: l = 0.5

  5. BM reinforcement model Partner’s Choice: C / D Aspiration Threshold Outcome [CC, DD, CD, DC] pa C / D Payoff [R, P, S, T] Stimulus D+C- T = 4 C+C+ R = 3 A = 2 D-D- P = 1 –1 <= Stimulus <= 1 S = 0 C-D+

  6. D+C- T = 4 C+C+ R = 3 A = 2 D-D- P = 1 S = 0 C-D+ BM reinforcement model Partner’s Choice: C / D Aspiration Threshold Outcome [CC, DD, CD, DC] pa C / D Payoff [R, P, S, T] Stimulus pD ↑↑ pC ↓↓ st(D / T) = 1 pC ↑ pC ↑ st(C / R) = 0.5 pD ↓ pC ↑ st(D / P) = –0.5 pC ↓↓ pC ↓↓ st(C / S) = – 1

  7. BM reinforcement model The Prisoner’s Dilemma pC= Probability to cooperate Aspiration Threshold: A = 2

  8. BM reinforcement model Partner’s Choice: C / D Aspiration Threshold Outcome [CC, DD, CD, DC] C / D Payoff [R, P, S, T] Stimulus pa If satisfactory,movetowards 1 a proportion (l·st) of the remaining distance If unsatisfactory, movetowards 0 a proportion (l·st) of the remaining distance

  9. BM reinforcement model 1 Learning rate: l = 0.5 pC C+ \ D- D+C- T = 4 C+ \ D- st = 1 C- \ D+ C+C+ R = 3 C- \ D+ st = 0.5 A = 2 C+ \ D- D-D- P = 1 st = -0.5 S = 0 C-D+ C- \ D+ st = -1 0 n= 0 n= 1 … n= 2

  10. BM reinforcement model The Prisoner’s Dilemma pC= Probability to cooperate Aspiration Threshold: A = 2

  11. D+C- T = 4 C+C+ R = 3 A = 2 D-D- P = 1 S = 0 C-D+ BM reinforcement model Most likely move Player 1 = D (T) Player 2 = C (S) Player 1 = C (R) Player 2 = C (R) Player 1 = D (P) Player 2 = D (P) Player 1 = C (S) Player 2 = D (T)

  12. D+C- T = 4 C+C+ R = 3 A = 2 D-D- P = 1 S = 0 C-D+ BM reinforcement model Expected motion

  13. MACY M W and Flache A (2002) Learning Dynamics in Social Dilemmas. Proc. Natl. Acad. Sci. USA 99(3), 7229-7236. Fixed aspirations, P < A < R Floating aspirations

  14. MACY M W and Flache A (2002) Learning Dynamics in Social Dilemmas. Proc. Natl. Acad. Sci. USA 99(3), 7229-7236. “We identify a dynamic solution concept, stochastic collusion, based on a random walk from a self-correcting equilibrium (SCE) to a self-reinforcing equilibrium (SRE). These concepts make much more precise predictions about the possible outcomes for repeated games.” “Rewards produce a SRE in which the equilibrium strategy is reinforced by the payoff, even if an alternative strategy has higher utility.” e.g.: PC1 = 1, PC2 = 1 with Ai < Ri Action/stimulus : C+C+; C+C+; C+C+; … PC1, PC2 : 1 1 ; 1 1 ; 1 1 ; …

  15. D+C- T = 4 C+C+ R = 3 A = 2 D-D- P = 1 S = 0 C-D+ BM reinforcement model Expected motion SRE

  16. MACY M W and Flache A (2002) Learning Dynamics in Social Dilemmas. Proc. Natl. Acad. Sci. USA 99(3), 7229-7236. “A mix of rewards and punishments can produce a SCE in which outcomes that punish cooperation or reward defection (causing the probability of cooperation to decrease) balance outcomes that reward cooperation or punish defection (causing the probability of cooperation to increase).” E(∆ PC) = 0 Prob 0.6 0.2 E(∆ PC) = 0.6 × 0.2 – 0.4 × 0.3 = 0 0.3 Prob 0.4 “The SRE is a black hole from which escape is impossible. In contrast, players are never permanently trapped in SCE”

  17. D+C- T = 4 C+C+ R = 3 A = 2 D-D- P = 1 S = 0 C-D+ BM reinforcement model Expected motion SRE SCE

  18. SRE: Challenges “The SRE is a black hole from which escape is impossible” “A chance sequence of fortuitous moves can lead both players into a self-reinforcing stochastic collusion. The fewer the number of coordinated moves needed to lock-in SRE, the better the chances.” However… The cooperative SRE implies PC = 1, but according to the BM model, PC = 1 can be approached, but not reached in finite time! How can you “lock-in” if the SRE cannot be reached? Floating-point errors? Can we give a precise definition of SREs?

  19. SCE: Challenges “The SCE obtains when the expected change of probabilities is zero and there is a positive probability of punishment as well as reward” E(∆ PC) = 0 But … Such SCEs are not always attractors of the actual dynamics.

  20. SCE: Challenges Such SCEs are not always attractors of the actual dynamics. E(∆ PC) = 0 SCE

  21. SCE: Challenges “The SCE obtains when the expected change of probabilities is zero and there is a positive probability of punishment as well as reward” E(∆ PC) = 0 But … Such SCEs are not always attractors of the actual dynamics! Apart from describing regularities in simulated processes, Can we provide a mathematical basis for the attractiveness of SCEs?

  22. Outline of the presentation • Motivation • In-depth analysis of the dynamics of the model • Formalisation of SRE and SCE • Different regimes in the dynamics of the system • Dynamics with HIGH learning rates (i.e. fast adaptation) • Dynamics with LOW learning rates (i.e. slow adaptation) • Validity of the expected motion approximation • Analysis of the robustness of the model • Conclusions

  23. A definition of SRE SRE: an absorbing state of the system CC CD 1 C+C+ C-D+ Space of states PC1 D+C- D+D+ 0 0 1 If Si < Ai < Pi DD DC PC2 Not an event Not an infinite chain of events SREs cannot be reached in finite time, but the probability distribution of the states concentrates around them

  24. A definition of SCE SCE of a system S: An asymptotically stable critical point of the continuous time limit approximation of its expected motion SRE SCE

  25. SRE & SCE E(∆ PC) = 0 NOT AN SCE SRE & SCE Expected movement of the system in a Stag Hunt game parameterized as [3 , 4 , 1 , 0 | 0.5 | 0.5 ]2. The numbered balls show the state of the system after the indicated number of iterations in a sample run.

  26. Three different dynamic regimes “By the ultralong run, we mean a period of time long enough for the asymptotic distribution to be a good description of the behavior of the system. The long run refers to the time span needed for the system to reach the vicinity of the first equilibrium in whose neighborhood it will linger for some time. We speak of the medium run as the time intermediate between the short run [i.e. initial conditions] and the long run, during which the adjustment to equilibrium is occurring.” (Binmore, Samuelson and Vaughan 1995, p. 10)

  27. By the ultralong run, we mean a period of time long enough for the asymptotic distribution to be a good description of the behavior of the system The long run refers to the time span needed for the system to reach the vicinity of the first equilibrium in whose neighborhood it will linger for some time. Ultralong run We speak of the medium run as the time intermediate between the short run [i.e. initial conditions] and the long run, during which the adjustment to equilibrium is occurring Long run Medium run Short run (initial conditions) Trajectories in the phase plane of the differential equation corresponding to the Prisoner’s Dilemma game parameterised as [ 4 , 3 , 1 , 0 | 2 | l ]2, together with a sample simulation run ( l=2−2 ). This system has a SCE at [ 0.37 , 0.37 ].

  28. The ultralong run (SRE) Most BM systems –in particular all the systems studied by Macy and Flace (2002) with fixed aspirations– converge to an SRE in the ultralong run if there exits at least one SRE.

  29. High learning rates (fast adaptation) Learning rate: l = 0.5 SRE The ultralong run(i.e. convergence to an SRE) is quickly reached. The other dynamic regimes are not clearly observed.

  30. Low learning rates (slow adaptation) Ultralong run Learning rate: l = 0.25 The three dynamic regimes are clearly observed. Long run Medium run -> trajectories Medium run Long run -> SCE Ultralong run -> SRE Short run (initial conditions)

  31. The medium run (trajectories) For sufficiently small learning rates and number of iterations n not too large (n·l bounded), the medium run dynamics of the system are best characterised by the trajectories in the phase plane. l = 0.3 n = 10 l = 0.03 n = 100 l = 0.003 n = 1000

  32. The long run (SCEs) When trajectories finish in an SCE, the system will approach the SCE and spend a significant amount of time in its neighbourhood if learning rates are low enough and the number of iterations n is large enough (and finite). This regime is the long run. Long run

  33. The ultralong run (SREs) Ultralong run Most BM systems – in particular all the systems studied by Macy and Flace (2002) with fixed aspirations– converge to a SRE in the ultralong run if there exits at least one SRE.

  34. l = 2-1 Learning rate l = 2-7 Iteration (time)

  35. The validity of mean field approximations The asymptotic (i.e. ultralong) behaviour of the BM model cannot be approximated using the continuous time limit version of its expected motion. Such an approximation can be valid over bounded time intervals but deteriorates as the time horizon increases. Ultralong run

  36. Outline of the presentation • Motivation • In-depth analysis of the dynamics of the model • Analysis of the robustness of the model • Model with occasional mistakes (trembling hands) • Renewed importance of SCEs • Discrimination among different SREs • Conclusions

  37. Model with trembling hands l = 0.25 noise = 0 l = 0.25 noise = 0.01 MODEL WITH TREMBLING HANDS ORIGINAL MODEL

  38. Model with trembling hands (no SREs) SREUP: SRE OF THE UNPERTURBED PROCESS The lower the noise, the higher the concentration around SREUPs.

  39. Model with trembling hands Importantly, not all the SREs of the unperturbed process are equally robust to noise. One representative run of the system parameterised as [ 4 , 3 , 1 , 0 | 0.5 | 0. 5 ] with initial state [ 0.9 , 0.9 ], and noise εi = ε = 0.1. Without noise: Prob( SRE[1,1] ) ≈ 0.7 Prob( SRE[0,0] ) ≈ 0.3

  40. Model with trembling hands Not all the SREs of the unperturbed process are equally robust to noise.

  41. Outline of the presentation • Motivation • In-depth analysis of the dynamics of the model • Analysis of the robustness of the model • Conclusions

  42. Conclusions • Formalisation of SRE and SCE • In-depth analysis of the dynamics of the model • Strongly dependent on speed of learning • Beware of mean field approximations • Analysis of the robustness of the model • Results change dramatically when small quantities of noise are added • Not all the SREs of the unperturbed process are equally robust to noise

  43. Reinforcement Learning Dynamics in Social Dilemmas Luis R. Izquierdo & Segis Izquierdo

More Related