Learning Dynamics for Mechanism Design An Experimental Comparison of Public Goods Mechanisms P.J. Healy [email protected] California Institute of Technology The Repeated Public Goods Implementation Problem • Example: Condo Association “special assessment” – Fixed set of agents regularly choosing public good levels. – Goal is to maximize efficiency across all periods – What mechanism should be used? • Questions: – Are the “one-shot” mechanisms the best solution to the repeated problem? – Can one simple learning model approximate behavior in a variety of games with different equilibrium properties? – Which existing mechanisms are most efficient in the dynamic setting? Previous Experiments on Public Goods Mechanisms I • Dominant Strategy (VCG) mechanism experiments – Attiyeh, Franciosi and Isaac ’00 – Kawagoe and Mori ’01 & ’99 pilot – Cason, Saijo, Sjostrom, & Yamato ’03 – Convergence to strict dominant strategies – Weakly dominated strategies are observed Previous Experiments on Public Goods Mechanisms II • Nash Equilibrium mechanisms – Voluntary Contribution experiments – Chen & Plott ’96 – Chen & Tang ’98 – Convergence iff supermodularity (stable equil.) • Results consistent with best response behavior A Simple Learning Model • k-period Best Response model – Agents best respond to pure strat. beliefs – Belief = unweighted average of the others’ strategies in the previous k periods • Needs convex strategy space – Rational behavior, inconsistent beliefs – Pure strategies only A Simple Learning Model: Predictions – Strictly dominated strategies: never played – Weakly dominated strategies: possible – Always converges in supermodular games – Stable/convergence => Nash equilibrium – Can be very unstable (cycles w/ equilibrium) A New Set of Experiments • New experiments over 5 public goods mechanisms – – – – – Voluntary Contribution Proportional Tax Groves-Ledyard Walker Continuous VCG (“cVCG”) with 2 parameters • Identical environment (endow., prefs., tech.) • 4 sessions each with 5 players for 50 periods • Computer Interface – History window & “What-If Scenario Analyzer” The Environment • Agents: i N N 5 • Private Good: xi Public Good: y Endowments: (i ,0) 2 • Preferences: ui ( xi , y ) bi y ai y xi i (ai , bi ) • Technology: y x / • Mechanisms: mi M i y(m) f (m1 , m2 ,, mn ) ti (m, y) gi (m1 , mn , y(m)) xi i ti (m, y) The Mechanisms y mi • Voluntary Contribution y mi • Proportional Tax • Groves-Ledyard • Walker • VCG ti iN y mi iN y mi ti iN M i R 2 iN y n ti ti mi y n y n 1 n 2 n mi i 2 2i mi 1 mi 1 y mi (aˆi , bˆi ) y yNPO (aˆ, bˆ) zi y NPO\{i} (aˆ , bˆ) n 1 2 2 ti y (b j y a j y ) (b j zi a j zi ) zi n j i j i Experimental Results I: Choosing k • Which value of k minimizes the M.A.D. across all mechanisms, sessions, players and periods? Model k=1 k=2 k=3 k=4 k=5 k=6 k=7 k=8 k=9 k=10 2-50 1.407 - 3-50 1.394 1.240 - 4-50 1.284 1.135 1.097 - 5-50 1.151 0.991 0.963 0.952 - • k=5 is the most accurate 6-50 1.104 0.967 0.940 0.932 0.924 - 7-50 1.088 0.949 0.925 0.915 0.9114 0.9106 - 8-50 1.072 0.932 0.904 0.898 0.895 0.897 0.899 - 9-50 1.054 0.922 0.888 0.877 0.876 0.881 0.884 0.884 - 10-50 1.054 0.913 0.883 0.866 0.860 0.868 0.873 0.874 0.879 - 11-50 1.049 0.910 0.875 0.861 0.853 0.854 0.863 0.864 0.870 0.875 Walker Session 2 Player 1 15 Message 10 5 0 -5 -10 0 10 20 30 40 50 40 50 Period Walker Session 2 Player 2 15 Message 10 5 0 -5 -10 0 10 20 30 Period Walker Session 2 Player 3 15 Message 10 5 0 -5 -10 0 10 20 30 40 50 40 50 Period Walker Session 2 Player 4 15 Message 10 5 0 -5 -10 0 10 20 30 Period Walker Session 2 Player 5 15 Message 10 5 0 -5 -10 0 10 20 30 40 50 40 50 Period Groves-Ledyard Session 1 Player 1 6 Message 4 2 0 -2 -4 0 10 20 30 Period Experimental Results: 5-B.R. vs. Equilibrium • Null Hypothesis: E[| m BR |] E[| m EQ |] t i t i t i t i • Non-stationarity => period-by-period tests • Non-normality of errors => non-parametric tests – Permutation test with 2,000 sample permutations • Problem: If EQit BRit then the test has little power • Solution: – Estimate test power as a function of ( EQit BRit ) / – Perform the test on the data only where power is sufficiently large. 0.9 0.95 0.8 0.94 0.7 0.93 0.6 0.92 0.5 0.91 0.4 0.89 0.3 0.86 0.2 0.8 0.1 0.67 0 0 0.5 1 1.5 2 2.5 ( - )/ a b 3 x 3.5 4 4.5 5 0 0 0.95 0 1 Prob. H False Given Reject H Estimated Test Power Simulated Test Power 5-period B.R. vs. Equilibrium • Voluntary Contribution (strict dom. strats): EQit BRit • Groves-Ledyard (stable Nash equil): EQit BRit • Walker (unstable Nash equil): 73/81 tests reject H0 – No apparent pattern of results across time • Proportional Tax: 16/19 tests reject H0 Interesting properties of the 2-parameter cVCG mechanism • Best response line in 2-dimensional strategy space Best Response in the cVCG mechanism Convert data to polar coordinates • i • Dom. Strat. = origin, B.R. line = 0-degree line s , ri s Experimental Results III: Efficiency • Outcomes are closest to Pareto optimal in cVCG – cVCG > GL ≥ PT > VC > WK (same for efficiency) – Sensitivity to parameter selection • Variance of outcomes: – cVCG is lowest, followed by Groves-Ledyard – Walker has highest • Walker mechanism performs very poorly – Efficiency below the endowment – Individual rationality violated 42% of last 10 periods Discussion & Conclusions • Data are consistent with the learning model. – Repercussions for theoretical research • Should worry about dynamics – k-period best response studied here, but other learning models may apply • Example: Instability of the Walker mechanism • cVCG mechanism can perform efficiently • Open questions: – cVCG behavior with stronger conflict between incentives and efficiency – Sensitivity of results to parameter changes – Effect of “What-If Scenario Analyzer” tool Efficiency Confidence Intervals - All 50 Periods Efficiency 1 No Pub Good 0.5 Walker VC PT Mechanism GL cVCG Av e rage Public Good Le v e ls 9 Pers 1-50 Public Good Level 8 Pareto Optimal Pers 41-50 7 6 5 4 3 2 1 0 VC PT GL WK Mechanism VCG VCG* Standard Deviation of PG Levels 7 Periods 1-50 Standard Deviation 6 Periods 41-50 5 4 3 2 1 0 VC PT GL WK Mechanism VCG VCG* Voluntary Contribution Mechanism Results
© Copyright 2026 Paperzz