37 min read

> "The best way to predict the future is to invent it." — Alan Kay

Chapter 41: The Frontier — Research Directions

"The best way to predict the future is to invent it." — Alan Kay

Prediction markets have moved from academic curiosities to operational platforms handling billions of dollars in notional volume. Yet the most exciting work lies ahead. Across a dozen subfields — from large language models that estimate probabilities to cryptographic protocols that protect trader privacy — researchers and practitioners are pushing the boundaries of what markets can do, who can participate, and what questions markets can answer.

This chapter is a guided tour of the research frontier. We will cover established lines of inquiry where significant progress has been made, emerging directions where the first results are appearing, and open problems where no one yet knows the answer. The goal is not merely to survey: for each direction, we provide enough technical depth — including working Python code — for you to start contributing.


41.1 The Research Landscape

41.1.1 Where We Stand

Prediction markets in 2025 sit at a remarkable inflection point. The core theory — proper scoring rules, logarithmic market scoring rules, automated market makers — is mature. Platforms like Polymarket, Metaculus, and Manifold Markets have demonstrated product-market fit. Regulatory frameworks are slowly catching up. But the gap between what markets could do and what they actually do remains enormous.

Consider just a few gaps:

Capability Current State Frontier Goal
Forecasting accuracy Good for high-liquidity political events Reliable for any well-defined question
Privacy Pseudonymous at best Cryptographically private
AI participation Manual traders with ad-hoc models AI agents as first-class market participants
Verification Requires trusted oracle Decentralized, manipulation-resistant
Scope Binary events, short horizons Conditional, causal, long-horizon questions
Accessibility Crypto-native users Universal, multi-chain, fiat-accessible

41.1.2 Academic vs. Industry Research

The research landscape splits along two axes:

Academic research focuses on theoretical guarantees — mechanism design with provable properties, information-theoretic bounds on aggregation, computational complexity of market equilibria. Key venues include EC (Economics and Computation), WINE (Web and Internet Economics), AAAI, NeurIPS, and specialized workshops.

Industry research focuses on practical scalability — gas-efficient AMM implementations, latency optimization, user experience, and regulatory compliance. Companies like Polymarket, Kalshi, and various DeFi protocols drive this work.

The most exciting developments happen at the intersection: ideas born in theory that become practical through engineering, and practical problems that demand new theory.

41.1.3 A Map of the Frontier

We organize research directions into five clusters:

  1. AI and forecasting: LLMs as forecasters, AI-augmented trading, automated strategy generation
  2. Privacy and security: Zero-knowledge proofs, homomorphic encryption, differential privacy
  3. Mechanism design: Information elicitation, peer prediction, automated mechanism design
  4. Causal and conditional markets: Markets for causal questions, interventional queries
  5. Infrastructure and applications: Cross-chain interoperability, novel applications

Each cluster has its own theoretical challenges, its own community of researchers, and its own timeline to practical impact. Let us dive in.


41.2 LLMs as Forecasters

41.2.1 The LLM Forecasting Revolution

Large language models have demonstrated surprising capability at probability estimation. When prompted appropriately, models like GPT-4, Claude, and Gemini can produce calibrated probability estimates for a wide range of questions — from geopolitical events to scientific outcomes.

The key insight is that LLMs have ingested vast amounts of text containing base rates, historical analogies, expert analyses, and probabilistic reasoning. When asked "What is the probability that X happens?", an LLM can synthesize this background knowledge in ways that sometimes rival or exceed human forecasters.

41.2.2 Benchmarking LLM Forecasts

The ForecastBench benchmark (Ye et al., 2024) provides a standardized evaluation framework. Questions are drawn from forecasting platforms and evaluated against resolutions. Key findings:

  • GPT-4 with chain-of-thought prompting achieves Brier scores competitive with the Metaculus community median on many question types.
  • Calibration varies by domain: LLMs are well-calibrated on geopolitical questions but poorly calibrated on long-tail scientific outcomes.
  • Recency bias is a major limitation: LLMs struggle with questions where the relevant information appeared after their training cutoff.
  • Overconfidence on questions with clear narratives: LLMs assign too-extreme probabilities when a compelling story is available.

The formal evaluation uses proper scoring rules. Recall the Brier score:

$$\text{Brier}(p, o) = (p - o)^2$$

where $p$ is the forecasted probability and $o \in \{0, 1\}$ is the outcome. A lower score is better. For $N$ questions:

$$\text{Brier}_{\text{avg}} = \frac{1}{N} \sum_{i=1}^{N} (p_i - o_i)^2$$

Calibration is assessed by grouping forecasts into bins and comparing the mean forecast to the mean outcome:

$$\text{Calibration Error} = \sum_{b=1}^{B} \frac{n_b}{N} |p_b - \bar{o}_b|$$

where $p_b$ is the mean forecast in bin $b$ and $\bar{o}_b$ is the mean outcome.

41.2.3 Prompt Engineering for Forecasting

The quality of LLM forecasts depends critically on prompting strategy. Research has identified several effective patterns:

Base Rate Prompting: Ask the LLM to first identify the reference class and base rate before adjusting for specific evidence.

What is the base rate for [event type]?
What factors in this specific case push the probability higher or lower?
Given these adjustments, what is your final probability estimate?

Adversarial Prompting: Ask the LLM to argue both sides before committing to a probability.

First, make the strongest case that X will happen.
Now, make the strongest case that X will NOT happen.
Weighing both arguments, what probability do you assign to X?

Decomposition Prompting: Break complex questions into conditionally independent sub-questions.

For X to happen, conditions A, B, and C must all hold.
P(A) = ?
P(B|A) = ?
P(C|A,B) = ?
P(X) = P(A) * P(B|A) * P(C|A,B) = ?

Calibration Anchoring: Provide the LLM with its own calibration statistics to improve self-awareness.

Historical analysis shows you tend to be overconfident when forecasting
[category]. Your 90% confidence intervals contain the true value only
75% of the time. Adjust accordingly.

41.2.4 Fine-Tuning for Prediction

Beyond prompting, researchers have explored fine-tuning LLMs on forecasting data:

  1. Supervised fine-tuning on (question, resolution) pairs from Metaculus and Good Judgment Open.
  2. RLHF with calibration reward: The reward signal incorporates proper scoring rules, penalizing miscalibration.
  3. Retrieval-augmented generation (RAG): Augmenting the LLM with real-time news and data to overcome the training cutoff.

The fine-tuning objective can be formulated as minimizing the expected Brier score:

$$\mathcal{L}(\theta) = \mathbb{E}_{(q, o) \sim \mathcal{D}} \left[ \left( f_\theta(q) - o \right)^2 \right]$$

where $f_\theta(q)$ is the model's probability estimate for question $q$ with parameters $\theta$, and $o$ is the binary outcome.

41.2.5 Current Limitations

Despite impressive results, LLM forecasters face fundamental challenges:

  1. Knowledge cutoff: LLMs cannot access information after their training date without external tools.
  2. Hallucination: LLMs may confidently cite nonexistent studies or statistics.
  3. Correlation blindness: LLMs struggle with questions where the answer depends on correlations between variables they have not seen together.
  4. Adversarial vulnerability: Carefully crafted questions can exploit LLM biases to produce systematically wrong forecasts.
  5. Lack of skin in the game: LLMs have no incentive to be accurate beyond their training objective. They cannot participate in markets to correct their own errors.

41.2.6 Python LLM Forecasting Harness

See code/example-01-llm-forecaster.py for a complete implementation of an LLM-based forecasting harness that: - Structures prompts using multiple strategies - Collects and aggregates probability estimates - Evaluates calibration against known outcomes - Compares LLM forecasts to market prices

# Quick preview of the LLM forecasting pattern
class LLMForecaster:
    def __init__(self, model_name: str, strategy: str = "decomposition"):
        self.model_name = model_name
        self.strategy = strategy

    def forecast(self, question: str, context: str = "") -> float:
        prompt = self._build_prompt(question, context)
        response = self._query_model(prompt)
        probability = self._extract_probability(response)
        return probability

    def evaluate(self, forecasts: list, outcomes: list) -> dict:
        brier = np.mean([(p - o)**2 for p, o in zip(forecasts, outcomes)])
        calibration = self._compute_calibration(forecasts, outcomes)
        return {"brier_score": brier, "calibration_error": calibration}

41.3 AI-Augmented Trading

41.3.1 The Human-AI Collaboration Spectrum

AI participation in prediction markets exists on a spectrum:

  1. AI as information source: Trader reads AI-generated analysis, makes own decision.
  2. AI as advisor: AI recommends trades, human approves.
  3. AI as co-pilot: AI executes routine trades, human handles edge cases.
  4. AI as autonomous agent: AI trades independently with allocated capital.
  5. AI as market maker: AI provides liquidity and sets prices.

Most current practice sits at levels 1-2. Research is pushing toward levels 3-5.

41.3.2 Combining Human and AI Forecasts

The optimal combination of human and AI forecasts is itself a prediction problem. The extremized average is a simple but effective method:

$$p_{\text{combined}} = \frac{p_H^{\alpha} \cdot p_A^{(1-\alpha)}}{p_H^{\alpha} \cdot p_A^{(1-\alpha)} + (1-p_H)^{\alpha} \cdot (1-p_A)^{(1-\alpha)}}$$

where $p_H$ is the human forecast, $p_A$ is the AI forecast, and $\alpha \in [0,1]$ controls the weighting.

Research from the Aggregative Contingent Estimation (ACE) program showed that the best aggregation methods: - Weight forecasters by track record - Apply recency weighting to account for information dynamics - Extremize the aggregate (push away from 50%) to account for shared information

For human-AI aggregation, additional considerations apply: - Correlation structure: Humans and AI may be correlated (both read the same news) or complementary (AI has better base rates, humans have better context). - Domain expertise: Weight the AI more heavily in domains where it has demonstrated calibration. - Confidence calibration: Adjust for the known miscalibration patterns of each source.

41.3.3 Reinforcement Learning for Market Making

Market making in prediction markets can be formulated as a reinforcement learning problem. The state includes the current order book, inventory, time to resolution, and any available information signals. The action is the bid-ask spread and position limits. The reward is trading profit minus inventory risk.

Formally, the MDP is:

  • State $s_t = (b_t, a_t, q_t, I_t, t)$ — bid, ask, inventory, information signal, time
  • Action $a_t = (\delta^b_t, \delta^a_t)$ — bid and ask offsets from midpoint
  • Transition $P(s_{t+1} | s_t, a_t)$ — determined by order flow model
  • Reward $r_t = \text{PnL}_t - \lambda \cdot q_t^2$ — profit minus inventory penalty

The Bellman equation is:

$$V(s) = \max_a \left[ r(s, a) + \gamma \sum_{s'} P(s'|s,a) V(s') \right]$$

Researchers have applied PPO (Proximal Policy Optimization) and SAC (Soft Actor-Critic) to learn market-making policies that outperform fixed-spread strategies in simulated environments.

41.3.4 Automated Strategy Generation

A more ambitious direction uses AI to generate trading strategies rather than just execute them. The process:

  1. Hypothesis generation: LLM generates candidate hypotheses about market inefficiencies.
  2. Feature engineering: LLM identifies relevant features from available data.
  3. Strategy coding: LLM writes backtestable trading code.
  4. Backtesting: Strategy is evaluated on historical data.
  5. Risk analysis: Strategy is stress-tested under adverse scenarios.
  6. Deployment: Approved strategies are executed with risk limits.

This pipeline is related to program synthesis in AI research. The key challenge is avoiding overfitting to historical data — a challenge that affects both human and AI strategy designers.

41.3.5 Python AI-Augmented Trader Prototype

The code/example-01-llm-forecaster.py file includes an AI-augmented trader class that demonstrates combining LLM forecasts with market data to generate trading signals.

class AIAugmentedTrader:
    def __init__(self, llm_forecaster, market_client, risk_params):
        self.forecaster = llm_forecaster
        self.market = market_client
        self.risk = risk_params

    def generate_signal(self, question, market_price):
        ai_prob = self.forecaster.forecast(question)
        edge = ai_prob - market_price
        if abs(edge) > self.risk.min_edge:
            size = self.kelly_size(ai_prob, market_price)
            return TradeSignal(direction=np.sign(edge), size=size)
        return None

41.4 Privacy-Preserving Prediction Markets

41.4.1 Why Privacy Matters

Privacy in prediction markets is not merely a convenience — it is essential for honest reporting. Consider the following scenarios:

  • A corporate insider wants to trade on a market about their company's quarterly earnings. Without privacy, their participation itself becomes a signal, and they face legal liability.
  • A government analyst wants to share their genuine assessment of a geopolitical event. Without anonymity, they fear career consequences for contradicting official positions.
  • A group of doctors want to aggregate their opinions on a medical outcome. Without privacy, peer pressure and reputation concerns distort their reports.
  • A trader with a successful strategy wants to trade without revealing their positions or strategy to competitors.

The fundamental tension is: markets aggregate information best when participants report honestly, but honest reporting requires protection from the consequences of that honesty.

41.4.2 Threat Models

To design privacy-preserving markets, we must first define what we are protecting against:

  1. Other traders: Cannot learn the identity, positions, or strategies of any specific trader.
  2. Market operator: Cannot link trades to real-world identities beyond what is necessary for compliance.
  3. External observers: Cannot determine who is participating or how.
  4. Governments: Cannot compel revelation of trader identities without proper legal process.

These threat models lead to different technical requirements:

Threat Protection Mechanism
Trade linkability Ring signatures, mixnets
Position revelation Homomorphic encryption
Identity disclosure Zero-knowledge proofs
Statistical inference Differential privacy

41.4.3 Zero-Knowledge Proofs for Anonymous Trading

A zero-knowledge proof (ZKP) allows a prover to convince a verifier that a statement is true without revealing anything beyond the statement's truth. In prediction markets, ZKPs enable:

  • Anonymous deposits: Prove you have sufficient funds without revealing your identity or balance.
  • Valid trade verification: Prove a trade is within your risk limits without revealing your current position.
  • Outcome verification: Prove that a claimed outcome is correct without revealing the verification process.

The mathematical framework uses the concept of a proof system $(P, V)$ where:

  • Completeness: If the statement is true, an honest prover can convince the verifier. $$\Pr[V(x, P(x, w)) = 1 \mid (x, w) \in R] = 1$$

  • Soundness: If the statement is false, no prover can convince the verifier (except with negligible probability). $$\Pr[V(x, P^*) = 1 \mid x \notin L] \leq \text{negl}(\lambda)$$

  • Zero-knowledge: The verifier learns nothing beyond the truth of the statement. $$\exists \text{Sim} : \text{View}_V[P(x,w) \leftrightarrow V(x)] \approx \text{Sim}(x)$$

Modern ZKP systems like zk-SNARKs (Zero-Knowledge Succinct Non-Interactive Arguments of Knowledge) make this practical. A trader can:

  1. Commit to a trade $t$ by publishing $\text{Com}(t) = g^t h^r$ (Pedersen commitment).
  2. Produce a proof $\pi$ that $t$ is valid (within balance, within risk limits).
  3. The market accepts the trade after verifying $\pi$, without ever learning $t$ directly.

41.4.4 Homomorphic Encryption for Hidden Order Books

Homomorphic encryption (HE) allows computation on encrypted data. Fully homomorphic encryption (FHE) supports arbitrary computations:

$$\text{Enc}(a) \oplus \text{Enc}(b) = \text{Enc}(a + b)$$ $$\text{Enc}(a) \otimes \text{Enc}(b) = \text{Enc}(a \cdot b)$$

For prediction markets, HE enables:

  • Encrypted order books: Orders are submitted encrypted. The matching engine operates on ciphertexts.
  • Private AMM interactions: The AMM computes prices and executes trades on encrypted positions.
  • Sealed-bid mechanisms: Traders submit encrypted bids; the mechanism computes the allocation without seeing individual bids.

The performance overhead of FHE remains the primary practical barrier. Current FHE schemes (CKKS, BFV, BGV) introduce latency of seconds to minutes for non-trivial computations. However, several optimizations are making FHE more practical:

  1. Batching: Process multiple trades in a single HE operation using SIMD-style packing.
  2. Hybrid schemes: Use HE only for the privacy-critical steps; use plaintext for the rest.
  3. Hardware acceleration: Custom ASICs and FPGAs for HE operations.

41.4.5 Secure Multi-Party Computation

Secure multi-party computation (MPC) allows $n$ parties to jointly compute a function $f(x_1, \ldots, x_n)$ without revealing their individual inputs $x_i$. For prediction markets:

  • Distributed price computation: Multiple operators jointly run the market without any single operator seeing all trades.
  • Threshold resolution: $k$-of-$n$ oracles must agree on an outcome, but no single oracle can be coerced.
  • Private aggregation: Forecasts are aggregated into a market price without revealing individual forecasts.

The Shamir secret sharing scheme divides a secret $s$ into $n$ shares such that any $k$ shares can reconstruct $s$ but fewer than $k$ shares reveal nothing:

$$f(x) = s + a_1 x + a_2 x^2 + \ldots + a_{k-1} x^{k-1}$$

Each party $i$ receives the share $f(i)$. Reconstruction uses Lagrange interpolation:

$$s = f(0) = \sum_{i \in S} f(i) \prod_{j \in S, j \neq i} \frac{j}{j - i}$$

41.4.6 Python Privacy Primitives

See code/example-02-privacy-primitives.py for implementations of: - Pedersen commitments for trade hiding - Simple ZKP for proving a value lies in a range - Shamir secret sharing for distributed market operation - Simulated homomorphic operations on encrypted order data


41.5 Differential Privacy for Market Data

41.5.1 The Problem of Market Data Publication

Prediction market operators want to publish aggregate statistics — volume, price history, order flow — to attract participants and demonstrate market health. But aggregate statistics can leak information about individual participants. A trader who places a large order at a specific time is identifiable from the price impact and volume spike.

Differential privacy provides a mathematical framework for quantifying and limiting this information leakage.

41.5.2 Differential Privacy Basics

A randomized mechanism $\mathcal{M}$ satisfies $(\epsilon, \delta)$-differential privacy if for any two neighboring datasets $D$ and $D'$ (differing in one record) and any set $S$ of possible outputs:

$$\Pr[\mathcal{M}(D) \in S] \leq e^{\epsilon} \cdot \Pr[\mathcal{M}(D') \in S] + \delta$$

The privacy parameter $\epsilon$ controls the privacy-utility tradeoff: - $\epsilon = 0$: Perfect privacy (but no useful output). - $\epsilon = 1$: Strong privacy. - $\epsilon = 10$: Weak privacy. - $\epsilon = \infty$: No privacy.

The key mechanisms are:

Laplace mechanism for numeric queries with sensitivity $\Delta f$:

$$\mathcal{M}(D) = f(D) + \text{Lap}\left(\frac{\Delta f}{\epsilon}\right)$$

where $\Delta f = \max_{D, D'} |f(D) - f(D')|$ is the global sensitivity — the maximum change in the query result from adding or removing one record.

Gaussian mechanism for $(\epsilon, \delta)$-DP:

$$\mathcal{M}(D) = f(D) + \mathcal{N}\left(0, \frac{2 \ln(1.25/\delta) \cdot \Delta f^2}{\epsilon^2}\right)$$

Exponential mechanism for non-numeric outputs: Select output $r$ with probability proportional to $\exp\left(\frac{\epsilon \cdot u(D, r)}{2\Delta u}\right)$, where $u$ is a utility function.

41.5.3 DP Mechanisms for Prediction Market Statistics

Applying DP to prediction market statistics requires careful analysis of sensitivity:

Price history: If the market uses a LMSR with liquidity parameter $b$, a single trader's maximum impact on the price is bounded by:

$$\Delta \text{price} \leq \frac{1}{b}$$

This gives us a natural sensitivity bound. Publishing DP price history requires adding noise calibrated to $1/b$.

Volume statistics: The sensitivity of total volume to a single trader is bounded by their position limit $q_{\max}$:

$$\Delta \text{volume} = q_{\max}$$

Order flow metrics: More complex statistics like order flow imbalance have higher sensitivity and require more noise.

41.5.4 Privacy-Utility Tradeoff

The fundamental challenge is that useful market data requires low noise, but privacy requires high noise. The tradeoff depends on:

  1. Market size: Larger markets tolerate more noise because the signal is stronger.
  2. Publication frequency: Real-time publication leaks more than periodic snapshots.
  3. Query complexity: Simple statistics (average price) require less noise than complex statistics (price impact curves).

The composition theorem governs how privacy degrades with multiple publications:

  • Basic composition: $k$ queries each with $\epsilon$-DP give $k\epsilon$-DP overall.
  • Advanced composition: $k$ queries give $(\epsilon\sqrt{2k\ln(1/\delta')} + k\epsilon(e^\epsilon - 1), k\delta + \delta')$-DP.

For market data published every minute over a day (1440 queries), basic composition gives a total budget of $1440\epsilon$. Advanced composition gives roughly $\sqrt{1440} \cdot \epsilon \approx 38\epsilon$, a significant improvement.

41.5.5 Python DP Implementation

See code/example-02-privacy-primitives.py for a DP market data publisher that: - Adds calibrated Laplace noise to price and volume statistics - Tracks cumulative privacy budget using the moments accountant - Demonstrates the privacy-utility tradeoff through simulation

class DPMarketPublisher:
    def __init__(self, epsilon_per_query: float, delta: float):
        self.epsilon_q = epsilon_per_query
        self.delta = delta
        self.queries_answered = 0

    def publish_price(self, true_price: float, sensitivity: float) -> float:
        noise = np.random.laplace(0, sensitivity / self.epsilon_q)
        self.queries_answered += 1
        return np.clip(true_price + noise, 0, 1)

    def total_privacy_spent(self) -> float:
        # Advanced composition
        k = self.queries_answered
        eps = self.epsilon_q
        return eps * np.sqrt(2 * k * np.log(1 / self.delta)) + k * eps * (np.exp(eps) - 1)

41.6 Information Elicitation Without Verification

41.6.1 The Verification Problem

Standard prediction markets require a trusted oracle to verify the outcome. But many important questions resist verification:

  • "Will AI be aligned by 2040?" — Who decides what "aligned" means, and when?
  • "Is this scientific paper's main result correct?" — Full verification may take years.
  • "What is the probability of an earthquake above magnitude 7 in the Bay Area in the next decade?" — We cannot wait a decade to resolve the market.
  • "How effective is this policy intervention?" — Counterfactuals are inherently unobservable.

For these questions, we need mechanisms that elicit honest reports without outcome verification.

41.6.2 Peer Prediction Mechanisms

Peer prediction mechanisms reward reporters based on the correlation between their reports and those of their peers, rather than on the truth. The core idea: if participants share a common prior and observe correlated signals, then honest reporting is a Bayesian Nash equilibrium of an appropriately designed scoring mechanism.

The Miller-Radzik-Prelec (MRP) mechanism works as follows:

  1. Each agent $i$ reports a signal $r_i$ and a prediction $\hat{p}_i$ of other agents' signals.
  2. Agent $i$ is scored based on: - An information score: How well $i$'s signal predicts a reference agent $j$'s signal. - A prediction score: How well $i$'s prediction of others' signals matches the actual distribution.

$$\text{Score}_i = S(r_j; \hat{p}_i) + \alpha \cdot S(r_i; \hat{p}_j) - S(r_j; \hat{p}_j)$$

where $S$ is a proper scoring rule (e.g., logarithmic) and $\alpha > 0$ is a weighting parameter.

The key theorem: Under the common prior assumption, honest reporting (reporting one's true signal and true posterior) is a strict Bayesian Nash equilibrium.

41.6.3 Bayesian Truth Serum

Prelec's Bayesian Truth Serum (BTS) is an elegant mechanism that incentivizes honest reporting of subjective judgments. Each agent:

  1. Reports an answer $r_i$ (e.g., "Yes" or "No").
  2. Reports a prediction $\hat{p}_i$ of the population distribution of answers.

The BTS score is:

$$\text{BTS}_i = \log\left(\frac{\bar{x}_{r_i}}{\hat{p}_{r_i}^{\text{geo}}}\right) + \alpha \cdot \text{PS}(\hat{p}_i, \bar{x})$$

where: - $\bar{x}_{r_i}$ is the empirical frequency of answer $r_i$ among other agents. - $\hat{p}_{r_i}^{\text{geo}}$ is the geometric mean of other agents' predictions for answer $r_i$. - $\text{PS}(\hat{p}_i, \bar{x})$ is a proper scoring rule comparing $i$'s prediction to the empirical distribution.

The first term is the information score: it rewards answers that are "surprisingly common" — more frequent in the sample than people predicted. The key insight is that truth-tellers produce answers that are surprisingly common, because they know the true state of the world and can correctly infer that others with similar evidence will give similar answers.

Formally: If agents share a common prior and observe conditionally independent signals, then each agent's posterior probability that others share their signal is higher than the population's prior prediction. Thus, the true answer will be "surprisingly common."

41.6.4 Limitations and Extensions

Peer prediction mechanisms face several limitations:

  1. Common prior assumption: If agents have different priors, honest reporting may not be an equilibrium.
  2. Collusion: If agents can coordinate, they can submit arbitrary reports that satisfy the mechanism's criteria.
  3. Small samples: With few agents, the empirical distribution is noisy, making scores unreliable.
  4. Multi-equilibrium problem: Besides honest reporting, there may be other equilibria (e.g., everyone reports the same thing).

Recent extensions address these issues:

  • Robust Bayesian Truth Serum (Witkowski and Parkes, 2012): Works with heterogeneous priors.
  • Peer Truth Serum (Radanovic and Faltings, 2013): Handles continuous signals.
  • Determinant-based Mutual Information (Kong and Schoenebeck, 2019): Information-theoretically optimal mechanism.
  • Surrogate scoring rules (Liu and Chen, 2023): Use "surrogate" outcomes derived from peer reports to replace actual outcomes in proper scoring rules.

41.6.5 Markets for Unverifiable Outcomes

Combining peer prediction with market mechanisms opens new possibilities:

  • Belief markets: Trade on the aggregate belief about an unverifiable proposition, scored using peer prediction.
  • Retrospective markets: Markets that resolve based on future expert consensus rather than objective verification.
  • Epistemic markets: Markets where the price represents the rational probability given available evidence, not the actual outcome.

These hybrid mechanisms are particularly relevant for AI safety — many key questions (e.g., "Will AI systems be controllable?") cannot be verified until it is too late.

41.6.6 Python Peer Prediction

See code/example-03-peer-prediction.py for implementations of the BTS and MRP mechanisms, including simulation of honest vs. strategic behavior under various conditions.


41.7 Automated Mechanism Design

41.7.1 The Mechanism Design Challenge

Traditional mechanism design derives optimal mechanisms through mathematical analysis — a process that requires significant expertise and often yields mechanisms that are optimal only under restrictive assumptions. Automated mechanism design (AMD) uses optimization and machine learning to discover good mechanisms directly.

The AMD framework: 1. Define the objective: What should the mechanism optimize? (e.g., information aggregation, liquidity, welfare) 2. Define the constraints: What properties must the mechanism satisfy? (e.g., incentive compatibility, budget balance, individual rationality) 3. Parameterize the mechanism space: Represent the class of possible mechanisms as a parameterized family. 4. Optimize: Search the mechanism space for the best mechanism using gradient descent, evolutionary algorithms, or reinforcement learning.

41.7.2 Learning Optimal AMMs

The automated market maker (AMM) is a natural target for AMD. Current AMMs (LMSR, CPMM) are hand-designed and may not be optimal. We can learn better AMMs by:

Step 1: Parameterize the cost function.

An AMM is defined by a cost function $C(\mathbf{q})$ where $\mathbf{q}$ is the vector of outstanding shares. The price of outcome $i$ is:

$$p_i = \frac{\partial C}{\partial q_i}$$

We parameterize $C$ as a neural network: $C_\theta(\mathbf{q})$.

Step 2: Define the objective.

We want the AMM to minimize the market maker's worst-case loss while maintaining price accuracy:

$$\min_\theta \max_{\mathbf{q}} \left[ C_\theta(\mathbf{q}) - C_\theta(\mathbf{0}) - \mathbf{q} \cdot \mathbf{o} \right]$$

subject to the constraint that $\nabla C_\theta(\mathbf{q})$ forms a valid probability distribution:

$$\sum_i \frac{\partial C_\theta}{\partial q_i} = 1, \quad \frac{\partial C_\theta}{\partial q_i} \geq 0$$

Step 3: Train via adversarial optimization.

Use gradient descent on $\theta$ and gradient ascent on $\mathbf{q}$ (minimax optimization):

$$\theta^{t+1} = \theta^t - \eta_\theta \nabla_\theta \mathcal{L}(\theta^t, \mathbf{q}^t)$$ $$\mathbf{q}^{t+1} = \mathbf{q}^t + \eta_q \nabla_\mathbf{q} \mathcal{L}(\theta^t, \mathbf{q}^t)$$

Early results show that learned AMMs can reduce worst-case loss by 10-30% compared to LMSR while maintaining comparable price accuracy.

41.7.3 Adaptive Mechanisms

A powerful extension is mechanisms that adapt to market conditions. An adaptive AMM adjusts its liquidity parameter based on observed trading patterns:

$$b_t = g_\phi(h_t)$$

where $h_t = (p_1, v_1, \ldots, p_t, v_t)$ is the history of prices and volumes, and $g_\phi$ is a learned function (e.g., an LSTM or transformer).

The objective is to minimize a combination of market maker loss and price tracking error:

$$\mathcal{L}(\phi) = \mathbb{E}\left[\text{Loss}(b_1, \ldots, b_T) + \lambda \sum_t (p_t - p_t^*)^2\right]$$

where $p_t^*$ is the "true" price (e.g., the price that would prevail with perfect information).

41.7.4 ML for Mechanism Design

Beyond AMMs, machine learning is being applied to broader mechanism design problems:

  • Auction design (Duetting et al., 2019): Neural networks learn revenue-optimal auctions.
  • Matching markets (Ravindranath et al., 2021): ML discovers better matching mechanisms.
  • Voting rules (Prasad et al., 2023): Optimization over parameterized voting rules finds mechanisms with desirable properties.

The connection to prediction markets: any mechanism that aggregates information from strategic agents faces similar design challenges. Techniques from ML-for-mechanism-design can be directly applied to design better prediction markets.

41.7.5 Python Mechanism Optimizer

See code/example-01-llm-forecaster.py (section on mechanism optimization) for a neural AMM training loop and an adaptive liquidity parameter learner.

class NeuralAMM:
    def __init__(self, n_outcomes: int, hidden_dim: int = 64):
        self.net = nn.Sequential(
            nn.Linear(n_outcomes, hidden_dim),
            nn.ReLU(),
            nn.Linear(hidden_dim, hidden_dim),
            nn.ReLU(),
            nn.Linear(hidden_dim, 1)
        )

    def cost(self, q: torch.Tensor) -> torch.Tensor:
        return self.net(q)

    def prices(self, q: torch.Tensor) -> torch.Tensor:
        q.requires_grad_(True)
        c = self.cost(q)
        return torch.autograd.grad(c, q, create_graph=True)[0]

41.8 Prediction Markets Meet Causal Inference

41.8.1 Markets for Causal Questions

Standard prediction markets answer observational questions: "What will happen?" But many important decisions require causal answers: "What would happen if we did X?"

A causal prediction market allows trading on conditional statements like:

  • "If the Fed raises rates by 50bp, what will inflation be in 12 months?"
  • "If drug A is approved, what will the 5-year survival rate for disease B be?"
  • "If policy X is implemented, what will the unemployment rate be?"

These are fundamentally different from observational conditional markets. The observational conditional $P(\text{inflation} | \text{Fed raises rates})$ includes selection effects: the Fed raises rates precisely when inflation expectations are high. The causal conditional $P(\text{inflation} | \text{do(Fed raises rates)})$ asks about the intervention, not the observation.

41.8.2 Potential Outcomes Framework

The potential outcomes framework (Rubin, 1974) provides the mathematical foundation. For each unit $i$ and each treatment $t \in \{0, 1\}$, there exists a potential outcome $Y_i(t)$. The causal effect is:

$$\tau_i = Y_i(1) - Y_i(0)$$

The fundamental problem of causal inference is that we can only observe one potential outcome for each unit. A causal prediction market addresses this by eliciting beliefs about both potential outcomes:

  • Market A: "What will Y be if treatment is applied?" $\to E[Y(1)]$
  • Market B: "What will Y be if treatment is not applied?" $\to E[Y(0)]$
  • Estimated causal effect: Price(A) $-$ Price(B) $\to E[\tau]$

The challenge is settlement: we can only observe one arm. Solutions include:

  1. Randomized resolution: Randomly select the treatment, then resolve the corresponding market. The other market is void.
  2. Decision markets: Markets that are designed to inform a decision, where the decision-maker commits to following the market's recommendation.
  3. Cross-population resolution: Use different populations or time periods for each arm.

41.8.3 Decision Markets

Decision markets (Hanson, 2013) are a special type of conditional prediction market where:

  1. There are markets for each possible decision: $P(Y | \text{do}(d))$ for each decision $d$.
  2. The decision-maker commits to choosing the decision whose market predicts the best outcome.
  3. Only the market corresponding to the chosen decision is resolved; others are voided.

The key result: if traders believe the decision-maker will follow the market, then in equilibrium, market prices reveal traders' true beliefs about causal effects. This is because the decision is now determined by the market prices, breaking the selection effect.

Formally, let $d^* = \arg\max_d E_{\text{market}}[Y | \text{do}(d)]$. If the decision-maker commits to $d^*$, then:

$$P_{\text{market}}(Y | \text{do}(d)) = P(Y | \text{do}(d))$$

for all $d$ in the neighborhood of $d^*$.

41.8.4 Causal Discovery from Market Data

An intriguing reverse direction: can we learn causal structure from prediction market data? If markets exist for many related variables, the pattern of price co-movement may reveal causal relationships.

Consider markets for events $A$, $B$, and $C$. If an exogenous shock to market $A$ causes a price change in market $B$ but not market $C$, this suggests $A \to B$ but $A \not\to C$.

Techniques from causal discovery — PC algorithm, GES, FCI — can be adapted to work with market price data. The key innovation is using the market's response to identifiable shocks as natural experiments.

41.8.5 Open Challenges

Causal prediction markets face several unresolved challenges:

  1. Thin markets: Conditional markets split liquidity. With $k$ possible decisions and $m$ possible outcomes, we need $k \times m$ markets, each potentially thin.
  2. Strategic manipulation: A trader who can influence the decision has an incentive to manipulate the market to cause their preferred decision.
  3. Common cause confounding: Market prices reflect both causal effects and common causes, making it difficult to isolate the causal component.
  4. Temporal consistency: Causal effects may change over time, but markets need stable enough prices to be informative.

41.9 Cross-Chain and Interoperable Markets

41.9.1 The Multi-Chain Reality

Prediction markets exist on multiple blockchains — Ethereum, Polygon, Solana, Gnosis Chain, and others. Each chain has its own advantages (speed, cost, liquidity, regulatory status) and its own user base. The fragmentation of liquidity across chains is a major obstacle to market efficiency.

41.9.2 Cross-Chain Bridging

Cross-chain bridges allow assets and messages to move between blockchains. For prediction markets, bridges enable:

  • Unified liquidity: Aggregate liquidity from multiple chains into a single virtual order book.
  • Chain-agnostic access: Trade on any market from any chain.
  • Arbitrage: Eliminate price discrepancies across chains.

The technical challenges are formidable:

  1. Security: Cross-chain bridges have been the target of major hacks (Ronin, Wormhole, Nomad). A bridge failure in a prediction market could result in incorrect resolutions or frozen funds.
  2. Latency: Cross-chain messages take minutes to finalize. This latency creates arbitrage opportunities that may be exploited by sophisticated actors.
  3. Consensus: The two chains may disagree on the state of the bridge. Resolving these disagreements requires careful protocol design.

41.9.3 Interoperability Protocols

Beyond bridges, deeper interoperability requires protocols that allow markets on different chains to interact:

  • Cross-chain order matching: An order on chain A can be matched with an order on chain B, with atomic settlement across both chains.
  • Shared oracle networks: Oracle networks (like Chainlink or UMA) provide the same resolution data to all chains.
  • Universal market identifiers: A standard naming scheme for markets that allows the same question to be traded across platforms.

The IBC (Inter-Blockchain Communication) protocol from the Cosmos ecosystem provides a template. IBC allows any two chains to exchange messages with guaranteed delivery and ordering. A prediction market protocol built on IBC could enable seamless cross-chain trading.

41.9.4 Layer 2 and Rollup Solutions

Layer 2 solutions — optimistic rollups (Arbitrum, Optimism) and ZK-rollups (zkSync, StarkNet) — offer a more practical near-term path to scalability:

  • Lower fees: Transaction costs of pennies rather than dollars.
  • Higher throughput: Thousands of transactions per second.
  • Inherited security: Settlement on Ethereum L1 provides security guarantees.

Prediction markets on L2 can serve retail users who are priced out of L1. The main challenge is liquidity fragmentation across L2s — a challenge that shared sequencers and cross-L2 messaging protocols are addressing.

41.9.5 The Vision: Universal Market Access

The ultimate goal is a world where any user, on any device, using any currency, can trade on any prediction market question. This requires:

  1. Chain abstraction: Users do not need to know which chain the market runs on.
  2. Gas abstraction: Users pay fees in their preferred currency.
  3. Intent-based trading: Users express their desired trade; a solver network finds the best execution across all available venues.
  4. Universal settlement: All trades settle on a shared, secure layer.

This vision is still years away, but the infrastructure is being built. Projects like Across Protocol, Connext, and Socket are building the cross-chain messaging layers that will make it possible.


41.10 Emerging Applications

41.10.1 AI Safety Prediction Markets

Prediction markets for AI safety represent one of the most consequential emerging applications. Key questions include:

  • "When will an AI system pass a comprehensive Turing test?"
  • "Will there be an AI-caused disaster resulting in >1000 deaths before 2035?"
  • "What is the probability that AI alignment is solved before AGI?"

The challenge: many of these questions involve tail risks with poorly defined base rates. Markets may be thin because few people have informed views. And the most important questions (e.g., existential risk) resist verification.

Research directions include: - Using peer prediction mechanisms for unverifiable AI safety questions. - Creating conditional markets: "If we invest $X in alignment research, what is the probability of success?" - Designing markets that aggregate the views of AI researchers specifically, weighted by expertise.

Organizations like AI Impacts and the Future of Humanity Institute have experimented with forecasting tournaments on AI timelines. Converting these to proper prediction markets could improve accuracy and provide real-time probability estimates.

41.10.2 Climate Adaptation Markets

Climate change creates enormous demand for probabilistic forecasts:

  • "What is the probability of a Category 5 hurricane hitting the Gulf Coast in 2026?"
  • "When will Arctic sea ice be ice-free in September?"
  • "What will the global mean temperature anomaly be in 2030?"

Climate adaptation markets could help: - Insurance pricing: Markets provide real-time risk estimates for catastrophe insurance. - Infrastructure planning: Municipal bonds tied to climate prediction markets signal where to invest in resilience. - Carbon credit pricing: Markets forecast the future price of carbon credits, enabling better hedging.

41.10.3 Scientific Replication Markets

The replication crisis has shown that many published scientific results do not replicate. Prediction markets on replication have proven surprisingly accurate:

  • Dreber et al. (2015) conducted replication markets for psychology studies. Markets predicted replication outcomes better than individual scientists.
  • Camerer et al. (2018) extended this to economics, with similar results.

Future directions: - Continuous replication markets: Rather than one-off studies, maintain persistent markets on the replicability of key results. - Pre-registration markets: Market on whether a pre-registered study will find the hypothesized effect. - Meta-scientific markets: Market on questions like "Will the p-value threshold be lowered to 0.005?"

41.10.4 Judicial Outcome Markets

Prediction markets on legal outcomes are contentious but potentially valuable:

  • "What is the probability that the Supreme Court upholds this law?"
  • "What will the verdict be in this trial?"
  • "What damages will be awarded in this class action?"

The primary concern is that such markets could compromise the integrity of the judicial process — a judge or juror who sees the market price might be influenced by it. On the other hand, existing markets (like litigation finance) already provide implicit probability estimates.

Research directions: - Post-verdict resolution: Markets resolve only after all appeals are exhausted. - Restricted participation: Only legal professionals can trade, reducing manipulation risk. - Anonymized questions: Market questions are phrased abstractly to avoid identifying specific cases.

41.10.5 Personal Prediction Markets for Self-Improvement

The most speculative application: prediction markets for personal development. Imagine:

  • A market on "Will I exercise 3+ times per week for the next month?" where you trade against a model of your own behavior.
  • A market on "Will I complete this project by the deadline?" that aggregates the opinions of your colleagues.
  • A market on "Will this habit change stick for 6 months?" that uses your historical data as the base rate.

The mechanism design challenge is maintaining incentive compatibility when the trader is also the subject of the prediction. Commitment devices (like Beeminder) are a crude version of this idea; full prediction markets could provide better calibration and more nuanced probability estimates.


41.11 Open Problems

The following open problems represent opportunities for researchers and practitioners. They are organized by domain and difficulty.

Theoretical Foundations

  1. Optimal aggregation under adversarial conditions: Design an aggregation mechanism that is optimal (in terms of information loss) even when a fraction $\alpha$ of participants are adversarial. Current results assume either all honest participants or a known corruption model. Difficulty: Hard.

  2. Tight bounds on market efficiency with bounded rational traders: What is the best-achievable price accuracy when traders have computational limitations? The current gap between rational-agent results and behavioral results is enormous. Difficulty: Hard.

  3. Mechanism design for correlated markets: When multiple markets share common information sources, how should mechanisms account for the correlation? Current AMMs treat markets as independent. Difficulty: Medium.

  4. Information-theoretic limits of prediction markets: What is the minimum number of traders needed to achieve a given level of price accuracy for a given information structure? Difficulty: Hard.

  5. Dynamic mechanism design for long-horizon markets: Markets that run for months or years face different incentive challenges than short-horizon markets. Characterize the optimal dynamic mechanism. Difficulty: Hard.

Privacy and Security

  1. Efficient ZKP for complex market operations: Current ZKP systems are too slow for real-time market operations. Design ZKPs that can verify complex trades (multi-leg, conditional) in under 100ms. Difficulty: Medium.

  2. Differential privacy with market microstructure: Extend DP theory to account for the specific structure of market data (order arrival times, price impact). Standard DP treats records as independent; market data is inherently sequential and correlated. Difficulty: Hard.

  3. Post-quantum privacy for prediction markets: Existing cryptographic primitives (used in ZKPs, HE) are vulnerable to quantum computers. Design quantum-resistant privacy-preserving market mechanisms. Difficulty: Hard.

  4. Collusion resistance in anonymous markets: When traders are anonymous, how can the mechanism detect and prevent collusion? This is fundamentally harder than collusion resistance in identified settings. Difficulty: Hard.

AI and Forecasting

  1. Calibration-optimal fine-tuning of LLMs: Design a fine-tuning procedure that minimizes calibration error (not just accuracy) for LLM forecasters. Current RLHF objectives do not directly optimize for calibration. Difficulty: Medium.

  2. LLM forecasting with formal uncertainty quantification: LLMs produce point estimates. Can we extract reliable confidence intervals? Ensemble methods and conformal prediction are promising directions. Difficulty: Medium.

  3. Adversarial robustness of AI forecasters: How sensitive are LLM forecasters to adversarial prompt manipulation? Can we certify robustness? Difficulty: Medium.

  4. Market-informed LLM training: Train LLMs using prediction market prices as training signal. The market provides a continuously-updated, well-calibrated probability for many events — a rich supervision signal. Difficulty: Medium.

  5. Multi-agent AI market simulation: Build a simulation environment where multiple AI agents trade in prediction markets. Study emergent behaviors: do AI markets converge to efficient prices? Do they exhibit new forms of manipulation? Difficulty: Medium.

Mechanism Design

  1. Learning optimal scoring rules from data: Instead of using hand-designed scoring rules (Brier, logarithmic), learn the scoring rule that maximizes information elicitation for a given population of forecasters. Difficulty: Medium.

  2. Peer prediction without the common prior: All existing peer prediction mechanisms assume some form of common prior. Can we design mechanisms that work under heterogeneous priors? Difficulty: Hard.

  3. Budget-balanced prediction markets: LMSR and other AMMs require a subsidizer. Can we design a mechanism that aggregates information as well as LMSR but is budget-balanced (expected subsidy of zero)? Difficulty: Hard.

  4. Combinatorial prediction markets at scale: Current combinatorial market mechanisms have exponential complexity in the number of events. Find a practical mechanism that scales to hundreds of correlated events. Difficulty: Hard.

Applications

  1. Prediction markets for reproducible science: Design a practical prediction market platform for scientific replication that is integrated into the publication process. Address incentive, liquidity, and ethical concerns. Difficulty: Medium.

  2. Real-time climate risk markets: Design markets that provide real-time probabilities for climate-related events, with proper handling of model risk and deep uncertainty. Difficulty: Medium.

  3. AI safety prediction markets with proper incentives: Design markets for long-term AI safety questions where the outcomes may not be verifiable for decades, if ever. Address the discounting problem and the verification problem. Difficulty: Hard.

  4. Personal prediction markets: Design a system where individuals can create prediction markets about their own lives, with proper incentive alignment when the subject is also a trader. Difficulty: Medium.

Infrastructure

  1. Gas-efficient on-chain AMMs: Reduce the gas cost of on-chain AMM operations by 10x through better data structures, batching, or L2-native designs. Difficulty: Medium.

  2. Cross-chain atomic settlement for prediction markets: Implement atomic settlement of prediction market trades across multiple chains without relying on trusted intermediaries. Difficulty: Hard.

  3. Decentralized oracle design with quantified reliability: Design an oracle mechanism where the probability of incorrect resolution is quantifiably bounded. Current oracles provide no formal guarantees. Difficulty: Hard.


41.12 Chapter Summary

This chapter has surveyed the frontier of prediction market research across five clusters:

  1. AI and forecasting: LLMs can produce calibrated probability estimates, especially with careful prompting and fine-tuning. AI-augmented trading combines human judgment with AI capabilities. Reinforcement learning shows promise for market making. The major open questions concern calibration, robustness, and the integration of AI agents as full market participants.

  2. Privacy and security: Zero-knowledge proofs, homomorphic encryption, and secure multi-party computation provide the cryptographic primitives for private prediction markets. Differential privacy allows publishing market statistics without revealing individual trades. The major open questions concern efficiency, scalability, and the privacy-utility tradeoff.

  3. Mechanism design: Peer prediction mechanisms enable information elicitation without outcome verification. Automated mechanism design uses optimization and ML to discover better market mechanisms. The major open questions concern the common prior assumption, budget balance, and scalability.

  4. Causal inference: Prediction markets can be adapted to answer causal questions through conditional markets and decision markets. Causal discovery from market data is an intriguing reverse direction. The major open questions concern thin markets, manipulation, and settlement.

  5. Infrastructure and applications: Cross-chain interoperability, L2 solutions, and universal market access are expanding the reach of prediction markets. Emerging applications include AI safety, climate adaptation, scientific replication, and personal prediction markets.

The field is vibrant and growing. The intersection of prediction markets with AI, cryptography, and causal inference is producing new questions faster than answers. This is the mark of a healthy research frontier.

Key Equations

Concept Equation
Brier score $\text{Brier}(p, o) = (p - o)^2$
Calibration error $\text{CE} = \sum_b \frac{n_b}{N} \|p_b - \bar{o}_b\|$
$(\epsilon,\delta)$-DP $\Pr[\mathcal{M}(D) \in S] \leq e^\epsilon \Pr[\mathcal{M}(D') \in S] + \delta$
Laplace mechanism $\mathcal{M}(D) = f(D) + \text{Lap}(\Delta f / \epsilon)$
Pedersen commitment $\text{Com}(v, r) = g^v h^r$
BTS score $\log(\bar{x}_{r_i} / \hat{p}_{r_i}^{\text{geo}}) + \alpha \cdot \text{PS}(\hat{p}_i, \bar{x})$
Neural AMM prices $p_i = \partial C_\theta / \partial q_i$

What's Next

In Chapter 42: Capstone Project, we bring together everything from the entire book. You will design, build, and evaluate a complete prediction market system — from mechanism design through implementation to empirical evaluation. The capstone integrates scoring rules, market making, privacy, and AI assistance into a single cohesive project that demonstrates mastery of prediction market theory and practice.


References

  • Ye, S., et al. (2024). "ForecastBench: A Dynamic Benchmark of AI Forecasting Capabilities." arXiv:2409.09839.
  • Prelec, D. (2004). "A Bayesian Truth Serum for Subjective Data." Science, 306(5695), 462-466.
  • Witkowski, J., and Parkes, D. C. (2012). "Peer Prediction without a Common Prior." EC 2012.
  • Kong, Y., and Schoenebeck, G. (2019). "An Information Theoretic Framework for Designing Information Elicitation Mechanisms That Reward Truth-telling." ACM Transactions on Economics and Computation, 7(1).
  • Hanson, R. (2013). "Shall We Vote on Values, But Bet on Beliefs?" Journal of Political Philosophy, 21(2), 151-178.
  • Duetting, P., et al. (2019). "Optimal Auctions through Deep Learning." ICML 2019.
  • Dreber, A., et al. (2015). "Using Prediction Markets to Estimate the Reproducibility of Scientific Research." PNAS, 112(50), 15343-15347.
  • Camerer, C., et al. (2018). "Evaluating the Replicability of Social Science Experiments in Nature and Science between 2010 and 2015." Nature Human Behaviour, 2, 637-644.
  • Gentry, C. (2009). "Fully Homomorphic Encryption Using Ideal Lattices." STOC 2009.
  • Dwork, C. (2006). "Differential Privacy." ICALP 2006.
  • Liu, Y., and Chen, Y. (2023). "Surrogate Scoring Rules." EC 2023.
  • Radanovic, G., and Faltings, B. (2013). "A Robust Bayesian Truth Serum for Non-Binary Signals." AAAI 2013.