Collective Revelation: A Mechanism for Self-Verified, Weighted, and Truthful Predictions Sharad Goel, Daniel M. Reeves, David M. Pennock Presented by: Nir Shabbat Outline • Introduction – Background and Related Work • The Settings • A Mechanism For Collective Revelation – The Basic Mechanism – A General Technique for Balancing Budgets • Summary Introduction • In many cases, a decision maker may seek to elicit and aggregate the opinions of multiple experts. • Ideally, would like a mechanism that is: 1. Incentive Compatible - Rewards participants to be truthful 2. Information Weighted - Adjusts for the fact that some experts are better informed than others 3. Self-Verifying - Works without the need for objective, “ground truth” observations 4. Budget Balanced - Makes no net transfers to agent Background and Related Work • Proper Scoring Rules – Rewards the agent by assessing his forecast against an actual observed outcome. • Agents are incentivized to be truthful. • For example, using the Brier scoring rule: We get: Background and Related Work • Peer Prediction Methods – Can be Incentive Compatible and Self-Verifying. • For example, the Bayesian truth serum asks agents to report both their own prediction and their prediction of other agents’ predictions. • Relies on the general property of informationscores, that a truthful answer constitutes the best guess about the most “surprisingly common” answer. Background and Related Work • With the appropriate reward structure , this framework leads to truthful equilibria: • 𝑥𝑘𝑟 = 1 iff response of agent 𝑟 is 𝑘 • 𝑦𝑘𝑟 - prediction of player 𝑟 for the frequency of 𝑘 • 𝑥𝑘 , 𝑦𝑘 - geometric average of predictions (or opinions) and meta predictions Background and Related Work • Predictions Markets – encourage truthful behavior and automatically aggregate predictions from agents with diverse information. • For example, a prediction market for U.S. presidential elections. Agents buy and sell assets tied to an eventual Democratic or Republican win. Each share pay $1 if the corresponding event occurs. Background and Related Work • Delphi Method – generates consensus predictions from experts generally through a process of structured discussion. • In each round, each expert anonymously provides his forecast and reasons, at the end of the round, market maker summarize the forecasts and reasons. • Process is repeated until forecasts converge. Background and Related Work • Competitive Forecasting – elicits confidence intervals on predictions, thereby facilitating information weighting. Like prediction markets, competitive forecasting rewards accuracy, though is not rigorously incentive compatible and relies on benchmarking against objective measurements. Background and Related Work Incentive Compatible Proper Scoring Rules ● Prediction Markets ○ Peer Prediction ○ Information Weighted ● ● Delphi Method Competitive Forecasting ● ○ ● Polls Collective Revelation SelfVerifying ● ○ ● ● The Settings 1. Common prior – Agents & Market Maker have a common prior on the distribution of 𝑋 2. Independent Private Evidence – Each agent 𝑖 privately observes 𝑛𝑖 ≥ 1 independent realization of the random variable 𝑋 (independent both of each other and across agents). 3. Rationality – Agents update their beliefs via Bayes’ rule. 4. Risk neutrality – Agents act to maximize their expected payoff. A Mechanism For Collective Revelation • Agents are scored against one another’s private information (Self Verification). • Agents report their subjective expectation of 𝑋, and their revised estimate in light of new hypothetical evidence (HFS). • From these 2 reports the subjective posterior, private information and also this agent’s prediction of other agents’ prediction is constructed. A Mechanism For Collective Revelation The Basic Mechanism LEMMA 1 1. 𝑋1 , … , 𝑋𝑛 𝑖. 𝑖. 𝑑 Bernoulli(𝑝) 2. An agent has subjective distribution 𝐹 for 𝑝 It is possible to elicit the kth moment of 𝐹 via a proper scoring rule that pays based on the outcome ( 𝑋1 , … , 𝑋𝑛 ) iff 𝑘 ≤ n The Basic Mechanism • Simply put, scoring against a single Bernoulli observation can only reveal the agent’s expectation and not the uncertainty of it’s prediction (as quantified by the variance). • So in order for a mechanism to be Information-Weighted (for Bernoulli outcome) it must score agents against multiple observations. The Basic Mechanism • For example, say we want to elicit an agent’s prediction regarding the distribution of 𝑋~Bernoulli(p). • We want to reward the agent using a proper scoring rule based on a single outcome 𝑋1 : 𝑈 𝑟 = 𝐶 − 𝑟 − 𝑋1 2 • To elicit the confidence, we must use another outcome 𝑋2 . For example, like this: 𝑈 𝑟, 𝑐 = 𝐶 − 𝑟 − 𝑋1 2 − 𝑐 − ∑ 𝑋𝑖 − 𝑋 2 2 The Basic Mechanism The Beta distribution is the “conjugate prior” of the Bernoulli distribution. That is, given 𝑋~Bernoulli(𝑝) with a prior p~Beta 𝛼, 𝛽 the posterior distribution of 𝑋, given 𝑛 independent trials out of which 𝑠 are successful, is: 𝐹 𝑝 𝑛, 𝑠 ~Beta 𝛼 + 𝑠, 𝛽 + 𝑛 − 𝑠 The Basic Mechanism LEMMA 2 • Let 𝑌 ~ Bernoulli 𝑝 • 𝑝~Beta 𝛼, 𝛽 a prior on 𝑝 • Fix 𝑛 > 0 (trials) and 0 ≤ 𝑠 ≤ 𝑛 (successes) integers • Define: That is: 𝑔𝑠,𝑛 𝛼 𝛼+𝑠 𝛼, 𝛽 = , 𝛼+𝛽 𝛼+𝛽+𝑛 The Basic Mechanism LEMMA 2 (Cont’d) Then: • 𝑔𝑠,𝑛 is a bijection from R + × R + to • And The Basic Mechanism LEMMA 3 • Let 𝑋1 , … , 𝑋𝑛 𝑖. 𝑖. 𝑑 Bernoulli 𝑝 (𝑛 ≥ 2) • Let 𝑝~Beta 𝛼, 𝛽 a prior on 𝑝 • Let 𝑆 = ∑𝑛𝑖=1 𝑋𝑖 • Define: The Basic Mechanism LEMMA 3 (Cont’d) Then: • ℎ𝑛 is a bijection from R + × R + to • And: The Basic Mechanism THEOREM 1 • Consider a Bayesian game according to the settings shown before with 𝑁 ≥ 3 players, 𝑋~Bernoulli(𝑝). • Let 𝑝~Beta 𝛼0 , 𝛽0 a common prior on 𝑝 • Let 𝐹𝑖 be the posterior dist. of agent 𝑖 for 𝑝 after updating according to its private info. • Fix 𝑛 > 0 and 0 ≤ 𝑠 ≤ 𝑛 integers The Basic Mechanism THEOREM 1 (Cont’d) • Suppose that each agent 𝑖 plays action 𝑎𝑖 , 𝑏𝑖 ∈ Ω where The Basic Mechanism THEOREM 1 (Cont’d) • Define: Where 𝑃𝑘 is projection into the kth component. The Basic Mechanism THEOREM 1 (Cont’d) • For arbitrary constants 𝐶𝑖 , set the reward of agent 𝑖 to be: The Basic Mechanism THEOREM 1 (Cont’d) • Then: (*) is a strict Nash Equilibrium. The Basic Mechanism PROOF • Fix attention on agent 𝑖, and suppose all agents 𝑗 ≠ 𝑖 play according to (*). −1 • By Lemma 2 - 𝑔𝑠,𝑛 𝑎𝑗 , 𝑏𝑗 = 𝛼𝑗 , 𝛽𝑗 parameters of 𝐹𝑗 . The Basic Mechanism PROOF (Cont’d) • Let 𝑠𝑗 , 𝑛𝑗 be private observations of agent 𝑗. Then: 𝛼𝑗 = 𝛼0 + 𝑠𝑗 𝛽𝑗 = 𝛽0 + 𝑛𝑗 − 𝑠𝑗 Consequently: The Basic Mechanism PROOF (Cont’d) • In particular, 𝑠−𝑖 = ∑𝑗≠𝑖 𝑠𝑗 and 𝑛−𝑖 = ∑𝑗≠𝑖 𝑛𝑗 • If 𝑖 plays according to (*) then: Since we are using the Brier Proper scoring rule, strategy (*) maximizes 𝑖’s expected reward. The Basic Mechanism PROOF (Cont’d) −1 • Moreover, since ℎ𝑛−1 ○ 𝑔𝑛,𝑠 is an injection, this is 𝑖’s unique best response, and so strategy (*) is a strict Nash equilibrium. The Basic Mechanism CORROLARY • Suppose all agents play the equilibrium strategy. • Define: The Basic Mechanism CORROLARY (Cont’d) • Define the aggregate information-weighted prediction as: Let 𝐹 𝑝 𝑋𝑖,𝑗 be the posterior dist. Resulting from the cumulative private evidence of all agents. Then 𝑝 = E𝐹 𝑋. A General Technique for Balancing Budgets • The idea is to reward agents via a Shared scoring rule. A simple application would be: And this is clearly budget balanced since ∑𝑛𝑖=1 𝑓 𝑟𝑖 , 𝑥𝑖 = 0 A General Technique for Balancing Budgets • In our setting, the “observations” 𝑥𝑖 are determined precisely by agents’ reports – simple application won’t work! • The solution proposed is to decouple scoring from benchmarking, e.g. agent 𝑖 is rewarded according to its performance relative to a set of agents whose score cannot be affected by agent 𝑖’s action. A General Technique for Balancing Budgets DEFINITION • Let 𝐺 be a Bayesian game with 𝑛 players • A 𝒌-projective family of 𝑮 is a family of games 𝐺𝑣 such that for each 𝑉 ⊆ 1, … , 𝑛 with 𝑉 = 𝑘, 𝐺𝑣 is a Bayesian game restricted to the players 𝑉 that preserves type spaces, action spaces, and players’ beliefs about types. In particular, 𝐺𝑣 is determined by a family of reward functions 𝑈𝒗 that specifies the expected reward for each player in V resulting from any given strategy profile of those players. A General Technique for Balancing Budgets THEOREM 2 • G is a Bayesian game with n players • 𝑟 = 𝑟1 , … , 𝑟𝑛 is a (strict) Nash equilibrium of G • Suppose 𝐺𝑣 is a k-projective family of G such that 2 ≤ 𝑘 < 𝑛/2 + 1 • Suppose that for each 𝑉 = 𝑗1 , … , 𝑗𝑘 , 𝑟𝑗1 , … , 𝑟𝑗𝑘 is a (strict) Nash equilibrium for 𝐺𝑣 . A General Technique for Balancing Budgets THEOREM 2 (Cont’d) • Then for any constant 𝐶 ∈ R, there are player rewards 𝑈𝑖 for the n-player game G such that: 1. For any strategy profile s, ∑𝑛𝑖=1 𝑈𝒊 𝑠 = 𝐶 2. The original equilibrium q is still a (strict) Nash equilibrium for the modified game In particular, by setting 𝐶 = 0, one can alter the rewards so that the game G is strongly budget balanced. A General Technique for Balancing Budgets PROOF • For 𝑖 ∈ 𝑉, denote: – 𝑈𝑖 - reward function of player 𝑖 in 𝐺 – 𝑈𝑉,𝒊 - reward function of player 𝑖 in the game 𝐺𝑉 – 𝑈𝑉,𝒊 𝑠 - reward of player 𝑖 in game 𝐺𝑉 , when 𝑠 is restricted to game G𝑉 . A General Technique for Balancing Budgets PROOF (Cont’d) • Denote the player sets Vi = 𝑖, 𝑖 + 1, … , 𝑖 + A General Technique for Balancing Budgets PROOF (Cont’d) • Summing over all players we get: A General Technique for Balancing Budgets PROOF (Cont’d) • Further more: And since 𝑘 < 𝑛/2 + 1, we have 2 𝑘 − 1 < 𝑛, and so 𝑖 ∉ 𝑉𝑖+𝑗 for 1 ≤ 𝑗 ≤ 𝑘 − 1. A General Technique for Balancing Budgets PROOF (Cont’d) • Consequently, if 𝑠𝑖 is the strategy of player 𝑖 and 𝑠−𝑖 is the strategy profile of all other players, then: Where 𝐶𝑖 is a function that does not depend on 𝑠𝑖 . A General Technique for Balancing Budgets PROOF (Cont’d) • Since (𝑟𝑖 , 𝑟−𝑖 ) is a (strict) Nash equilibrium for 𝐺𝑉𝑖 : And so, using the new reward functions, the original equilibrium 𝑟 is still a (strict) Nash equilibrium. Summary • We’ve seen a mechanism, i.e. Collective Revelation, for aggregating experts opinions which is: – Incentive Compatible – Information Weighed – Self-Verifying • We’ve seen a general technique for constructing budget balanced mechanisms that applies both to collective revelation and to past peerpredictions method. THE END
© Copyright 2026 Paperzz