Modern mechanism des.. - Carnegie Mellon School of Computer

Mechanism design for
computationally limited agents
(previous slide deck discussed the case where valuation determination was
complex)
Tuomas Sandholm
Computer Science Department
Carnegie Mellon University
Part I
Mechanisms that are computationally
(worst-case) hard to manipulate
Voting mechanisms that are worst-case hard to
manipulate
• Bartholdi, Tovey, and Trick. 1989. The computational
difficulty of manipulating an election, Social Choice and
Welfare, 1989.
• Bartholdi and Orlin. Single Transferable Vote Resists
Strategic Voting, Social Choice and Welfare, 1991.
• Conitzer, V., Sandholm, T., Lange, J. 2007. When are
elections with few candidates hard to manipulate? JACM.
• Conitzer, V. and Sandholm, T. 2003. Universal Voting
Protocol Tweaks to Make Manipulation Hard.
International Joint Conference on Artificial Intelligence
(IJCAI).
• Elkin & Lipmaa …
2nd-chance mechanism
[in paper “Computationally Feasible VCG Mechanisms” by Nisan & Ronen, EC-00, JAIR]
• (Interesting unrelated fact in the paper that has had lots of follow-on
research: Any VCG mechanism that is maximal in range is IC)
• Observation: only way an agent can improve its utility in a VCG
mechanism where an approximation algorithm is used is by helping the
algorithm find a higher-welfare allocation
• Second-chance mechanism: let each agent i submit a valuation fn vi
and an appeal fn li: V->V. Mechanism (using alg k) computes k(v),
k(li(v)), k(l2(v)), … and picks the among those the allocation that
maximizes welfare. Pricing based on unappealed v.
Other mechanisms that are worst-case
hard to manipulate
• O’Connell and Stearns. 2000. Polynomial Time
Mechanisms for Collective Decision Making,
SUNYA-CS-00-1
• …
Part II
Usual-case hardness of manipulation
Impossibility of usual-case hardness
•
For voting:
– Procaccia & Rosenschein JAIR-97
• Assumes constant number of candidates
• Impossibility of avg-case hardness for Junta
distributions (that seem hard)
– Conizer & Sandholm AAAI-06
• Any voting rule, any number of candidates,
weighted voters, coalitional manipulation
• Thm. <voting rule, instance distribution> cannot be
usually hard to manipulate if
–
It is weakly monotone (either c2 does not win, or if
everyone ranks c2 first and c1 last then c1 does not
win), and
– If there exists a manipulation by the manipulators,
then with high probability the manipulators can only
decide between two candidates
– Elections can be Manipulated Often by Friedgut,
Kalai Nisan FOCS-08
• For 3 candidates
• Shows that randomly selected manipulations work
with non-vanishing probability
– Isaksson, Kindler&Mossel FOCS-10
• For more than 3 candidates
– Still open directions available
• Multi-stage voting protocols
• Combining randomization and manipulation
hardness…
•
Open for other settings
Problems with mechanisms that are
worst-case hard to manipulate
• Worst-case hardness does not imply
hardness in practice
• If agents cannot find a manipulation, they
might still not tell the truth
– One solution avenue: Mechanisms like the one
in Part III of this slide deck...
Part III
Based on “Computational Criticisms
of the Revelation Principle” by
Conitzer & Sandholm
Criticizing truthful mechanisms
• Theorem. There are settings where:
– Executing the optimal truthful (in terms of social welfare)
mechanism is NP-complete
– There exists an insincere mechanism, where
• The center only carries out polynomial computation
• Finding a beneficial insincere revelation is NP-complete for the agents
• If the agents manage to find the beneficial insincere revelation, the
insincere mechanism is just as good as the optimal truthful one
• Otherwise, the insincere mechanism is strictly better (in terms of s.w.)
• Holds both for dominant strategies and Bayes-Nash
implementation
Proof (in story form)
• k of the n employees are needed for a project
• Head of organization must decide, taking into account
preferences of two additional parties:
– Head of recruiting
– Job manager for the project
• Some employees are “old friends”:
• Head of recruiting prefers at least one pair of old friends on
team (utility 2)
• Job manager prefers no old friends on team (utility 1)
• Job manager sometimes (not always) has private
information on exactly which k would make good team
(utility 3)
– (n choose k) + 1 types for job manager (uniform distribution)
Proof (in story form)…
Recruiting: +2 utility for pair of friends
Job manager: +1 utility for no pair of friends, +3 for the
exactly right team (if exists)
• Claim: if job manager reports specific team preference,
must give that team in optimal truthful mechanism
• Claim: if job manager reports no team preference, optimal
truthful mechanism must give team without old friends to
the job manager (if possible)
– Otherwise job manager would be better off reporting type
corresponding to such a team
• Thus, mechanism must find independent set of k
employees, which is NP-complete
Proof (in story form)…
Recruiting: +2 utility for pair of friends
Job manager: +1 utility for no pair of friends, +3 for the
exactly right team (if exists)
• Alternative (insincere!) mechanism:
– If job manager reports specific team preference, give that team
– Otherwise, give team with at least one pair of friends
• Easy to execute
• To manipulate, job manager needs to solve (NP-complete)
independent set problem
– If job manager succeeds (or no manipulation exists), get same
outcome as best truthful mechanism
– Otherwise, get strictly better outcome
Criticizing truthful mechanisms…
• Suppose utilities can only be computed by
(sometimes costly) queries to oracle
u(t, o)?
u(t, o) = 3
oracle
• Then get similar theorem:
– Using insincere mechanism, can shift burden of
exponential number of costly queries to agent
– If agent fails to make all those queries, outcome can
only get better
Is there a systematic approach?
• Previous result is for very specific setting
• How do we take such computational issues into account
in general in mechanism design?
• What is the correct tradeoff?
– Cautious: make sure that computationally unbounded agents
would not make mechanism worse than best truthful
mechanism (like previous result)
– Aggressive: take a risk and assume agents are probably
somewhat bounded
• Recent results on these manipulation-optimal
mechanisms in [Othman & Sandholm SAGT-09]