The Prediction Markets We Were Promised: A Guide to Slaying Dragons With Oracles, Commitments, and Blockchains

In this post we'll walk through the inner-workings of prediction markets and the systems they exist inside of.

The Prediction Markets We Were Promised: A Guide to Slaying Dragons With Oracles, Commitments, and Blockchains

Prediction markets have circulated as a promising application of blockchains since at least 2013, where they were briefly mentioned in the Ethereum whitepaper. Shortly after, in 2014, proposals for decentralized prediction markets re-emerged concurrently in two orthogonal architectural lineages—systems run by automated market makers (AMMs) that provide always-on pricing and liquidity, and systems powered by order-book markets that rely on explicit bids and asks. Ever since, the ecosystem has produced a long tail of variants, hybrids, and reinventions (see, e.g., Appendix A here for a list).

Despite this line of work, the dominant paradigm adopted by today's prominent prediction market platforms have been converging towards increasingly more centralized, more permissioned, and more trust-based systems. In the most popular platforms (e.g., Polymarket and Kalshi), the systems' critical levers—market creation, liquidity provision, outcome resolution, and settlement—remain subject to human discretion or governance that can be influenced after positions are established and exploited during market operation at users' expense.

The real-world manifestations of this "killer crypto-native dApp" we've talked about for over a decade have seemingly hit the mainstream, but they've completely diverged from the core thesis that made pairing prediction markets with blockchains so enticing. If prediction markets are meant to be effective mechanisms that users can wield to aggregate information from disparate sources, then this trend away from the decentralized prediction markets we were promised isn't a design detail—it’s a dead end.

In this post we'll walk through the inner-workings of prediction markets and the systems they exist inside of. We'll show that building truly decentralized, permissionless, and trustless prediction markets is full of dragons. Our goal is to chart precisely where these dragons live, why existing platforms keep sailing around them (or have fallen prey to them), and why we may finally be able to slay them due to a confluence of recent technological advances.

This post kicks off our journey, but the final destination is still on the horizon. Over the coming days we'll release a series of posts that tell the whole picture. We've split the story into five pieces:

  • Part I: What Are Prediction Markets?
    Here we define prediction markets and the problems they're designed to solve. Our introduction is kept at a high-level, and only digs into technical details that'll help understand or build intuition for discussions later on. Beyond abstractions, we also highlight certain facets of prediction market platforms that are critical for understanding how prediction markets exist in reality.
  • Part II: The On-Chain Prediction Market Thesis
    Prediction markets have been in the collective imagination of people building blockchain technologies for over a decade. In this part I try to unpack why on-chain prediction markets are so enticing, and what properties we need to ensure for the vision to hold.
  • Part III: Praxis and the Illusion of Progress
    Popular prediction market platforms already exist, and are increasingly entering mainstream discourse. Are we on a trajectory that leads back to the thesis we envision? Or are we converging on something else? From the title of this part, you can imagine I argue the latter. The goal of isn't to criticize any individual platform; it's to identify the dragons we need to face and the traps we need to avoid when building prediction market platforms.
  • Part IV: Survival and a Need To Redefine Monetization
    From the perspective of making revenue, the model adopted by prominent platforms today is "successful". However, is it aligned with supporting accurate prediction markets? I argue here that the answer is "No". It's aligned with supporting prediction markets that maximize volume and volatility. We discuss an alternative approach, wherein platforms stop behaving like venues and instead monetize as infrastructure.
  • Part V: Returning to the Thesis
    Finally, after a long journey, we circle back to the original question. How do we build the prediction markets we were promised? In this part I illustrate how a set modern tools—scalable chains, oracle games, verification protocols, and AI—could be brought together to make the thesis a reality. Discussions are kept at a high-level (honestly, each subsection could've easily filled its own series of posts...) and the claim is not that the path forward is easy. Instead, I argue that we finally have the necessary ingredients and can reasonably start building platforms aligned with prediction markets we were promised. A strategy, some grit, and the resolve to innovate is the final barrier in our way.

Part I:
What Are Prediction Markets?

Suppose you (the aggregator) would like to obtain a prediction about an uncertain variable, and you suspect there exist individuals (informants) out in the world who hold private signals bearing on the eventual outcome. How should you go about eliciting and aggregating relevant signals from disparate informants into a single coherent prediction? How should informants reason about revealing their high-quality, and perhaps only partial, information? How do we incentivize strategic informants to reveal private information honestly? These questions have been studied extensively in the field of mechanism design under the umbrella of information aggregation problems, and prediction markets are designed to solve these problems by making "being correct" tradeable.

Informally, prediction markets work as follows. The aggregator defines one or more contingent claims (i.e., securities, contracts, or assets) whose payoffs depend on the realized outcome being predicted. Informants then trade these claims however they please. When the prediction market is well designed, each trade has a concrete interpretation as an implicit report of the trader's belief and the mechanism yields a collective forecast the aggregator can read off directly.

The remainder of this part is dedicated to laying some groundwork we'll need later on. Readers who primarily want a broad-strokes understanding of our arguments, or who've engaged with formal treatments of prediction markets, can skip to Part II without too many worries. Our discussion is tailored towards setting the stage and providing slightly more formal intuition that could help.

Specifically, below we'll define prediction markets a bit more rigorously, introduce common market microstructures (i.e., the trading/price-formation mechanism), and highlight some key platform-level decisions required for these markets to function.

Arrow-Debreu Securities

For the sake of clarity, and to keep discussions concrete, much of this series will focus (sometimes implicitly, sometimes explicitly) on prediction markets for Arrow–Debreu (AD) securities. This is the canonical setting users of today's popular prediction market platforms will typically have in mind when reasoning about what prediction markets do: Contracts pay 1 if a particular outcome occurs and 0 otherwise, so the market price can be read like a probability over outcomes. With that said, although some specific technical details may need to be massaged or extended, the overarching arguments we'll make throughout the post can be applied to more general classes of prediction markets.

Let \(\mathcal{Y} = \{y_1, \dots, y_n\}\) be the set of mutually exclusive outcomes the event could take, and let \(Y \in \mathcal{Y}\) denote the realized outcome. A prediction market for AD securities has \(n\) assets \(A_1, \dots, A_n\) that informants can buy and sell, where each share of \(A_i\) pays 1 unit of currency if \(Y=y_i\) and 0 otherwise. If an informant’s belief assigns probability \(q_i = \Pr[Y=y_i]\), then in an idealized setting (no discounting, no frictions, etc.) their expected value of one share of \(A_i\) is \(q_i\) and their "fair price" for \(A_i\) should therefore be \(p_i = q_i\). Hence, assuming traders make decisions accordingly, the current market prices \((p_1,\dots,p_n)\) are naturally interpretable as a collective probability distribution over \(\mathcal{Y}\).

From this understanding of how "price \(\approx\) probability" we're able to derive why market incentives turn trading into a channel for directly reporting beliefs when informants are strategic. If the market currently prices \(A_i\) at \(p_i\) and you believe the true probability is \(q_i\), then buying one additional share has expected profit \(q_i - p_i\) (and selling has expected profit \(p_i - q_i\)). In idealized settings a trader's best response strategy reduces to simply: buy when the market price is below your belief, sell when it is above, and stop when prices match your belief. Obviously, in practice, prices will deviate from traders' exact probabilities due to complexities such as risk aversion, fees, etc. that'll inevitably exist. Our goal shouldn't be to design systems that match these theoretical ideals exactly, but to build systems that aspire towards them; to that end, I believe a well-designed prediction market makes moving prices toward your belief the natural profit-seeking action rather than outmaneuvering other traders or exploiting systemic inefficiencies.

A Concrete Example:
In a competition with \(n\) competitors, \(\mathcal{Y}\) could be the set \(\{\)"Competitor 1 will win"\(, \dots,\) "Competitor \(n\) will win"\(\}\) and \(Y\) will take on a value in \(\mathcal{Y}\) corresponding to who ultimately won the competition. For each \(i \in \{1,\dots, n\}\), informants buy or sell shares of \(A_i\) depending on their belief about what \(Y\) will be. At the end of the competition, when it's been determined that competitor \(i\) won, we know \(Y=\)"Competitor \(i\) will win" is true so each share of \(A_i\) pays out 1 unit of currency whereas each \(A_{j \neq i}\) pays out 0. In the end, informants with well-calibrated beliefs are expected to profit.

Market Microstructures

Prediction markets aren't a monolith, they're better thought of as a family of mechanisms whose behaviours can vary dramatically depending on the market's microstructure—i.e., how the prices are set, liquidity is provided, and trades are executed. For the purposes of this series we'll compartmentalize these design decisions into two archetypes:

1) Cost-function Prediction Markets
A cost-function prediction market (CFPM) is driven by a convex cost function \(C:\mathbb{R}^n\to\mathbb{R}\) derived from a market scoring rule, and is equivalent to popular classes of AMMs used in DeFi. CFPMs maintain a market state \(x \in \mathbb{R}^n\), where \(x_i\) is the net outstanding shares of \(A_i\). If a trader wants to buy/sell a bundle \(\Delta \in \mathbb{R}^n\), their "cost" to move the market from state \(x\) to \(x+\Delta\) is given by \(\text{cost}(\Delta\mid x) \;=\; C(x+\Delta) - C(x)\). The instantaneous market prices are given by the gradient \(p(x) = (p_1(x), \dots, p_n(x)) = \nabla C(x)\), which (for appropriately chosen functions \(C\)) provably live on the probability simplex (i.e., \(p_i(x) \ge 0\) and \(\sum_i p_i(x)=1\)). As a result, CFPMs ensure you have a continuously defined probability vector (and hence well-defined prices) at all times—crucially, even when no one is trading.

The relationship between CFPMs and proper scoring rules also guarantee that, under standard economic assumptions, traders' expected profit is maximized by trading until prices match their true belief. That is, the mechanism makes "moving the market toward what you believe" the locally optimal behavior while simultaneously ensuring your target always exists and is well-defined.

A Canonical Example:
Perhaps the best known CFPM is the Logarithmic Market Scoring Rule (LMSR), where \[C(x) \;=\; b \log\Big(\sum_{j=1}^n e^{x_j/b}\Big), \quad\text{so that}\quad p_i(x) \;=\; \frac{e^{x_i/b}}{\sum_{j=1}^n e^{x_j/b}}.\] The parameter \(b>0\) controls liquidity/price responsiveness, where larger \(b\) makes prices move more slowly (deeper liquidity) but increases the market maker’s worst-case loss. In particular, for LMSR, the standard bounded-loss guarantee is simply \(b\log n\).

2) Order Matching Prediction Markets
We use the umbrella term order matching prediction markets (OMPMs) to describe a broad family of standard microstructures; those where the market mechanism must match each trader with a counterparty, and prices are inferred from the matched orders. These mechanisms are the engine behind today's most popular prediction market platforms (e.g., Polymarket and Kalshi), and the best known member of the family is the central limit order book (CLOB).

Instead of a fixed pricing rule, OMPMs use a collection of standing offers:

  • A trader posts a limit order: "Buy \(A_i\) at price \(p_i \le \pi\)" or "Sell \(A_i\) at price \(p_i \ge \pi\)."
  • The system matches compatible orders; trades occur when bids meet asks.

Generally, at any given moment, OMPMs imply a piecewise pricing curve from the visible depth, but it does not provide the same kind of global, always-on probability vector we saw in CFPMs. For example:

  • If there are no orders, there may be no price.
  • Even with orders there may be a wide spread, and the "price" becomes a microstructure choice (last trade, mid-price, best bid, best ask, etc.) made by the platform. Hence "prices", and therefore collective probabilities, may no longer be straightforward to interpret.
  • Liquidity can vanish instantly if participants cancel or pull orders, so the mechanism has no built-in "always-on" commitment.

As we'll discuss later, for these reasons and others, trading strategy in OMPMs becomes entangled with timing, other traders’ behavior, and adverse selection.

Examples in "The Wild":
Under the hood, both Polymarket and Kalshi use CLOBs. Some technical details differ between the two platforms, but both also take a centralized approach to matching limit orders.

Surrounding Market Ecosystem

Until now we’ve been treating prediction markets as abstract information aggregators—you define contingent claims, let informants trade, and read the collective forecast off of the prices. Although incredibly useful, in practice, fixating only on the abstraction hides where many dragons reside. A prediction market deployed in practice must bridge two worlds: The clean internal world where the market's mechanics unfold (trades, allocation of shares, price updates, etc.), and the messy external world where events resolve and predictions are interpreted (politics, scientific results, court rulings, etc.).

The moment these worlds collide, hard design questions emerge at their intersection. Nuanced complexity lives where mechanism design's rigor must meld with any imprecision in the domains they're applied to—at the interface between logic and natural language, between off-chain reality and on-chain enforcement, and between how traders interact with a system and externalities it's coupled with.

Microstructure is one of these design decisions, but we discussed it separately as it directly impacts market participants' behaviours and the platforms' decentralization. However, beyond microstructure, we note three platform-level decisions that shape whether markets can remain aligned with the trustworthy information aggregators we were promised during difficult edge cases—or whether they warp into discretionary betting products riddled with "escape hatches" for admins.

Market Topics

Topics specify what markets are trying to predict; in the AD securities framing, it defines the outcome set \(\mathcal{Y}\). Although intuitive at a high-level, natural language is full of edge cases and complexities when paired with context. Questions of the form "Will \(E\) happen by date \(D\)?" highlight this phenomenon well: What counts as "happen", which time zone matters, what data source is authoritative, and what happens if the world produces something adjacent-but-not-identical (e.g., "wins" vs "concedes", "dies" vs "declared dead"). When moving prediction markets from theory to practice, topic design requires products to intermingle with mechanisms; ambiguity is not just cosmetic, and changes incentives while inviting adversarial trading strategies.

Recent Examples:
As we've seen throughout 2025 on Polymarket, these linguistic edge cases not only arise in practice but also take on numerous forms. The "Zelenskyy suit" market left traders in arms about what defines a "suit". The "Astronomer divorce" market presumed the involved couples were married (vs, e.g., common law), and required additional context be added later. Markets whose outcome space need to be mapped to a form convenient for a platform (e.g., scalar predictions are often bucketized to match the "Yes-No" share structure on platforms like Polymarket and Kalshi) can warp trader incentives when their belief doesn't align with buckets and can introduce artificial volatility (e.g., the "April temperature increase" market).

Resolution Mechanisms

Resolution mechanisms determine which claims can be redeemed at settlement; in terms of AD securities, these mechanisms decide which outcome \(Y=y_i\) hold. For some topics resolution is trivial, e.g. an on-chain invariant or a deterministic computation. However, for many topics, resolution reduces to a social choice problem that needs to untangle a web of subjectivity, credibility, and misaligned incentives. A resolution mechanism has to define who can report outcomes, how reports are verified, how disputes work, how long finality takes, and what the system does when reality is genuinely contested. In other words: the mechanism needs to establish a trusted "oracle" that remains credible even when large positions are at stake and adversarial pressures arise. When trust in resolution mechanisms crumble, it impacts not only the contested market but can also indirectly affect how traders will approach future markets.

Recent Examples:
Again, just in 2025, examples abound on Polymarket. The "Ukrain mineral deal before April" market's resolution is widely contested and has resulted in allegations that their resolution mechanisms were manipulated. Similarly, the "Fordow nuclear facility destruction" market remains contested as the involved countries reported differing outcomes and no neutral third party was granted access.

3) Commitment Protocols
Predictions are often contingent on a very specific set of priors, and commitment protocols serve to give traders "alignment" on those priors by defining what is locked once a market goes live—topic wording, data sources, dispute rules, settlement rules, fee schedules, discretionary powers held by humans or governance, etc. For prediction markets to serve as well-designed information aggregators in practice, commitments are not optional niceties; they are a market’s cornerstone. Traders will calibrate their beliefs subject to these commitments, and these commitments have deep implications on how effectively information can be injected into markets. How an event is interpreted, the integrity of the mechanism that adjudicates it, etc. all inform trading strategies and, in turn, the information contained in each trade. If critical terms can change after positions are established, then the "market price" is no longer a clean forecast; it is entangled with governance risk, admin risk, and meta-speculation about how the platform will behave under stress. Commitment protocols are often hidden under other platform-level decisions (or altogether overlooked) in discussions about prediction markets, but end-to-end commitments need to be made explicit in these complex systems or else rational actors will seek out unexpected ways to trade accuracy for profit seeking.

Examples in "The Wild":
Continuing with our examples on Polymarket, their commitment protocols are essentially folded into various design decisions and relatively opaque. A market's rules are posted with the market page's creation, but almost every facet of the market can be "clarified" (aka "changed" or "amended") right up until the final resolution. In some cases, the tradable outcomes in a market can change mid-flight or they can all be effectively nullified (e.g., if the determined outcome is unknown/50-50) at settlement. Even the resolution mechanisms, which are committed to in theory, rely on human discretion (from UMA token holders, evidence submitters, etc.) and are so malleable that they can be influenced while markets are live. The lack of explicit and well-defined commitments gives rise to a class of mechanism-aware trading strategies that aim to explicitly capitalize on these uncertainties rather than introducing information to markets, e.g. the "dispute alpha" strategy.

In Conclusion...

We introduced prediction markets with a skew towards future discussions in this post. This is intentional, and our introduction is by no means comprehensive. Prediction markets have been studied extensively and, depending on your focus, can be compartmentalized into a much more granular framework. For readers interested in diving into this rabbit hole, I suggest the following:

Alternatively, although out of scope for this post so I left it out, wagering mechanisms are a very related class of mechanisms that're generally under-represented in conversations about aggregate forecasting. These mechanisms have shared ancestry with CFPMs, but are built with slightly different assumptions about the aggregator and how informants want to participate. I'll probably write a seperate post at some point about wagering mechanisms and forecasting tournaments, but this paper can serve as a nice introduction to the topic although not being a survey per se.

In what's coming, I'll leverage the ideas and tools we highlighted here to analyze real world prediction markets and the platforms they exist inside of. As many readers will have noted, the platform-level decisions we've introduced are tightly coupled—a topic is only as good as the resolution mechanism that can adjudicate it, and a resolution mechanism is only as credible as the commitments preventing it from being swapped or pressured ex post. The inter-relation between these decisions is precisely where many dragons live, and is a focal point of our upcoming discussions.

Part II:
The On-Chain Prediction Market Thesis

When prediction markets first entered the blockchain technology zeitgeist in the early 2010's, they promised to be much more than some variant of "a decentralized gambling app". They embodied something closer to a public utility:
A mechanism that turns earnest belief into a collective forecast—without requiring anyone to trust the operator.

The Markets We Were Promised

The Ethereum whitepaper states the core intuition behind this thesis quite casually; with an oracle (or a oracle-like Schelling mechanism), prediction markets are "easy to implement" and might even become a mainstream substrate for futarchy.  The claim here is not that prediction markets are trivial, but that the blockchain provides the missing enforcement layer—collateral can be escrowed, payouts can be executed mechanically, and participants don’t need to grant discretionary power to an intermediary.

A decade later, the most prominent prediction markets have indeed shown flashes of that promise, and Vitalik’s more recent "info finance" framing captures the significance. His argument boils down to something along these lines:

Prediction markets don’t just serve the traders placing bets, they can serve everyone who reads the charts. Markets become a unique kind of informational public good—a way to compress scattered signals into a continuously updated, legible forecast. Crucially, the value is not limited to a handful of headline events; the long-term opportunity is a world of "micro" questions—myriad small, targeted markets designed to extract specific facts, judgments, or conditional predictions—if the underlying substrate and participants can support it.

The thesis was never merely to "put these markets on a blockchain" or "allow people to gamble on future events". It was to use blockchains to make prediction markets credible by construction. To enable markets that anyone can access, that no one can rewrite at their discretion, and that verifiably settle according to rules that were committed to before the money arrived.

This thesis is easy to get behind in the abstract, but once you begin writing it down as concrete technical requirements you begin to see the silhouette of dragons rising on the horizon. Avoiding these dragons is why so many attempts retreat into permissioned market creation, discretionary resolution, governance-supported backdoors, and non-commital liquidity provisioning. These dragons are not philosophical; they are implementation details that bite, but I believe they're surmountable with modern techniques and conviction (more on this in Part V).

The Properties We Should Demand

If we want to build prediction markets that serve as the public utility we've envisioned—and not as a permissioned version designed around the goals of platforms themselves, their liquidity providers, etc.—then decentralized prediction market platforms need to satisfy a set of axiomatic properties and make strong, principled commitments. These are not for aesthetics, but rather they're invariants needed to keep the system honest when incentives turn adversarial, edge cases arise, and circumstances no longer serve specific stakeholders.

1) Permissionless market creation, with enforceable topic guardrails
A central part of the promise is that "anybody can create a market", but it is critical that we're realistic that "anything can be a market topic" is an ethical and legal minefield. For a platform to credibly support permissionless market creation it needs a way to support open creation while constraining the space of allowable topics such that it doesn’t reintroduce human vetoes as a central point of failure. In practice, this means markets must be specified in a algorithmically-verifiable format (clear outcome set, timeframes, data sources, etc.) and topic restrictions must be enforced by protocol-level constraints (or, at least, by constraints that are themselves credibly neutral and not manipulable ex post).

2) Always-on liquidity with committed capital
A prediction market that has no price for long stretches of time is not aggregating information; it is intermittently hosting bids and asks. Ensuring coherent probabilities throughout the entirety of a market's lifespan requires an always-on pricing mechanism and a liquidity commitment structure that prevents rational liquidity from evaporating precisely when the market becomes most informative (near resolution, under controversy, etc.). The design goal doesn't equate to having deep liquidity, but rather to ensuring liquidity cannot disappear (strategically, arbitrarily, or otherwise) after other participants have relied on it.

3) Prices moved by beliefs, not microstructure meta-strategies
A well-designed and efficient prediction market should make being right the reliable strategy for making money. In other words, the marginal P&L of a trade should be driven mainly by the gap between your belief and the current implied probability—not by your ability to harvest spreads, rebates, queue position, latency, or temporary price voids. Importantly, this doesn't mean eliminating arbitrage in the broad sense; some arbitrage is healthy as it enforces logical coherence across related markets (e.g., "A or B" vs "A" and "B"). Our target here is much narrower, we want markets to facilitate a contest of research and calibrated beliefs rather than rewarding strategies based on timing, execution, and structural quirks.

4) Precommitted resolution that is reproducible and tamper-resistant
All roads lead to resolution in prediction markets, and resolution mechanisms can color or even kill these markets in practice. If the resolution mechanism can be swapped, influenced, or "interpreted" after positions are established, then trading ultimately reduces to governance speculation. To reliably aggregate signals throughout a market's duration, a platform must commit—at market creation—to a resolution mechanism that is: specified in advance (including dispute handling and tie-breakers), reproducible (third parties can independently replay and verify the outcome), and credibly resistant to manipulation by the market creator, the platform operator, or any "council" that can be captured.

5) Automatic settlement triggered by the resolution mechanism
Settlement is the end of the causal chain cementing a market's incentives. If resolution produces an outcome but settlement requires human approval (signing a transaction, blessing a payout, etc.), then the platform has reintroduced a trusted intermediary at the final step. The property we need is simple: resolution implies settlement. Once the precommitted resolution mechanism finalizes, redemption should be directly and automatically enabled (or executed) by the same rules that were locked in when the market was born.

6) Credible neutrality over the full lifecycle
Even if all these properties are satisfied on paper, decentralization fails if the platform can censor participation, delist markets, change fee schedules, pause trading, or unilaterally rewrite contracts through upgrade keys. These may sometimes be defensible product choices, but they are also governance levers—and governance levers become attack surfaces when money is on the line. Once a market is created, its terms are final and sacred; it must run its course until predefined finality conditions are satisfied, and without exception.

In Conclusion...

Taken together, these properties serve as a north star that guide us towards the prediction markets we were promised: An always-on probability layer that can’t be leaned on by insiders, that can’t be rewritten by governance once started, and that settles by construction rather than by permission.

These properties, and building the ideals in this thesis, underlie the analyses and arguments we make throughout this post. In particular, our next part will look at how existing prediction market platforms measure up against these properties.

Part III:
Praxis and the Illusion of Progress

On the surface it's easy to look at the last few years of prediction markets and assume that "the thesis is working". Compared to earlier platforms (e.g., Augur) the interfaces are smoother and the liquidity is thicker in flagship markets. Charts from platforms like Polymarket and Kalshi have broken into the mainstream; they circulate on social media while being cited by politicians and pundits in the same breath as polls. A decade ago this would have sounded like fiction. However, most of the progress towards the thesis is exactly that—surface-level smoke and mirrors. While the UI has improved, the underlying infrastructure—what blockchains were supposed to make credible by construction—has gradually centralized.

In this section we’ll take a step back and look at modern platforms and how they measure up against the properties we previously set forth. We’ll look at where they concentrate control—market creation, liquidity provision, resolution, settlement—and how those concentrations predictably surface as inefficiencies, fragility near resolution, and exploitable opportunities whenever disputes are likely.

Whose Discretion Matters After Positions Are Established?

The "control surface" of prediction market systems in practice can be understood along four axes:

  • Market creation \(\rightarrow\) Who can define what is being predicted? Under what Constraints?
  • Liquidity \(\rightarrow\) Who ensures prices exist? What stops liquidity from disappearing when inconvenient for liquidity providers?
  • Resolution \(\rightarrow\) Who decides what the outcome is? Using what evidence? What are the rules for disputes? What guarantees do we have?
  • Settlement \(\rightarrow\) Who triggers or enable payouts? Can anyone block, delay, or rewrite it?

Platforms can call themselves "decentralized" while still centralizing along any of these axes, but if any facet within (or even between) these is mutable ex post—i.e., once money arrives and there is skin in the game—then traders are no longer making decisions based purely on their beliefs about what'll happen in the world. Instead, trading is (and rationally should be) driven by beliefs about a mixture of reality and platform behaviour under stress. The market prices in such a system cease to be the clean forecasts promised by the thesis; they become a noisy composite signal contaminated by meta-speculation about how the system behaves at critical junctures.

Given the intensity of these statements, I want to make two points clear. Firstly, modern platforms absolutely do aggregate information and can yield useful forecasts. Secondly, what follows is not an indictment of any single platform. Rather, I'm making claims about design attractors that I believe existing platforms are being pulled towards and that ultimately are self-defeating if our goal is to build towards the thesis. The uncomfortable truth is that today’s prominent platforms are not early versions of what'll lead us to the "info finance" ideals that drive the zeitgeist. They are, increasingly, a different product category—one that converges toward gambling-like incentives and systemic choices favoring specific stakeholders while siphoning credibility from a well-intentioned thesis.

A Diagnostic Lens: Looking Along the Axes of Control

To make this analysis concrete, it'll help to briefly reframe the properties we demanded previously as operational questions:

  1. Can anyone start a market without permission—and are topic guardrails enforced by protocol or human veto?
  2. Are probability vectors well-defined throughout the market's lifespan, and can liquidity be withdrawn once others have relied on it?
  3. Are prices being moved by traders' beliefs about the realized outcome, or by their understanding of the platform's behaviours?
  4. Is resolution explicitly precommitted, reproducible, verifiable, and tamper-resistant?
  5. Does resolution mechanically imply settlement, without a final discretionary signature?
  6. Is the system credibly neutral over the full lifecycle of the market, or can it censor, pause, delist, amend, or rewrite?

Many modern platforms provide compelling answers to some of these questions—especially when it improves UX, reduces operational risk, or maximizes profits. That said, answering the questions that matter most for enshrining credibility by construction are usually left as an afterthought and are where centralization creeps in. We'll see why as we follow the four control axes that platforms tend to design around.

Market Creation: Permissionless in Marketing, Permissioned in Protocols

A truly open substrate for prediction markets is structurally incompatible with "predict anything" if it's meant to exist in the real world. This isn't a moral stance, it's a systemic reality. Allowing unconstrained creation of market topics invites markets that are unethical, illegal, coercive, or violent—markets whose existence creates negative externalities even if they never settle. Any platform that wants to exist within (and preferably contribute to) society must confront this.

Ultimately, prediction markets are tools for information aggregation and, like all tools, are not appropriate for all jobs. Hence keeping our trajectory aligned with the thesis doesn't require ensuring "can anyone create a market on any topic"—it's requires solving "how do we design robust systems that allow anyone to create a market within a class of topics we're able to support". The distinction between these two problems is often conflated and misrepresented in standard talking points about prediction markets, and designing products to maintain the illusion that your platform supports "markets on any topic" will culminate in fundamentally different technologies. The complexities of scoping and enforcing market topic guardrails is the first dragon that platforms often confront, and beware because it's one that'll rear its ugly head throughout our discussions below.

With this intuition in mind, the bifurcating decision becomes how topic constraints are enforced: Protocol-enforced topic guardrails are automated constraint systems that are legible, checkable, and binding; they restrict what can be expressed (or at least what can be instantiated) without depending on any operator's discretion after the fact. By contrast, platform-enforced topic guardrails require discretionary review by some trusted party and, even if the trusted party is well-intentioned, the existence of a veto is an inherently centralizing force.

The most prominent platforms almost always choose platform-enforced guardrails, e.g. market creation on Polymarket and Kalshi. They curate markets and narrow what can be created by relying on internal policy, compliance review, and editorial judgment. That choice is understandable—and often legally necessary—when the goal is to present yourself as enabling "markets on any topic". Unfortunately, this choice also has deep downstream consequences, as the platform effectively becomes both the editor and the publisher of reality's questions; it can no longer serve as a neutral substrate.

In this newfound dual role the platform becomes a bottleneck and must optimize accordingly to survive. The goal is no longer to support and facilitate markets that maximize informational value. It necessarily becomes to optimize for attention, retention, controversy, repeat trading, etc. while simultaneously ensuring markets remain within operational and legal risk thresholds.

Once the system is built around platform-enforced topic guardrails its center of gravity shifts by construction. The informational public utility we were promised wants a long tail of micro-markets and niche questions, whereas the platform businesses that dominate today need a manageable catalog of high-volume and high-engagement markets. Historically, attempts have been made at building platforms with protocol-enforced topic guardrails (or lack of guardrails), but the class of topics they support are either extremely narrow (e.g., Thales) or again "markets on any topic" (which runs into rather predictable pitfalls) .

Interestingly, the dragon we see here is surprisingly straightforward to rout if we confront it methodically and with conviction. Sailing around this dragon by resorting to platform-enforced guardrails will leave your platform ill-equipped when it confronts subsequent threats. Confronting it recklessly, without appropriate defenses, by naively scoping the class of markets your platform can handle will kill your platform outright or leave it debilitated. In either case, the lesson is simple: A platform without appropriate scope and guardrails for market creation can easily start looking less like "an information aggregator for the people" and more like either "a den of thieves and assassins" or "a bookie for whatever is currently legible, tradable, and safe enough to host".

Liquidity: Available When Convenient, Absent Otherwise

Beyond market creation we must confront questions about liquidity, where a platform's true commitments are revealed.

In our post introducing prediction markets, we highlighted two popular families of market microstructures: cost function prediction markets (CFPMs) and order matching prediction markets (OMPMs).

In CFPMs, liquidity isn't something you hope for; it's contractually cemented at market creation. Prices exist because the mechanism provably guarantees they exist, and if you want to move probabilities then you can always do so at a well-defined cost. Critically, once capital is committed in a CFPM, it serves as a market parameter that cannot evaporate when conditions become adverse.

Despite these strong liquidity guarantees, CFPMs are not used by popular prediction market platforms today—OMPMs are undisputably the dominant microstructure choice. These platforms use CLOBs where bids and asks appear and disappear at the will of participants (or a small set of professional market makers). Despite their popularity, order matching has two immediate side-effects: Prices aren't guaranteed to exist (especially in thin markets) and can have unintuitive probabilistic interpretations. Liquidity is least reliable when adverse selection spikes because rational liquidity providers widen spreads, reduce size, or pull orders.

The adoption of OMPMs on these platforms is not a consequence of any high-minded principle or technical superiority; in fact, it'd fair to say that CFPMs have stronger economic guarantees and more rigorous underpinnings. The ubiquity of OMPMs is almost certainly rooted in their lack of risk for platforms while simultaneously appealing to professional market makers whom platforms (with their newfound role as "publishers") want to attract. At the end of the day, microstructure is not an implementation detail. It's a decision about who the market rewards and who takes on risk.

As we've discussed at length, a prediction market mechanism that reliably aggregates information should reward traders for being right about the world. Yet profit opportunities in OMPMs often come from being right about market flow; anticipating other traders’ timing, harvesting spreads, exploiting latency, or positioning around liquidity vacuums.

To be clear, I'm not claiming that professional market making is evil or has no place. I also believe OMPMs can be well-suited microstructure choices for specific markets where sufficiently high-volume, high-engagement, tight spreads, etc. are reasonable assumptions. However, in general, prediction market platforms accommodate a long-tail of thin markets, which are precisely the settings where reliance on a small set of actors who are not committed—actors who can enter and exit at will—results in markets that become least liquid precisely when they're most informative. In other words, these platforms become venues whose informational capacity is contingent on voluntary, profit-maximizing liquidity that can leave whenever the venue becomes stressful.

Liquidity without commitment is a mirage, and a prediction market without liquidity is a hollow facade. Confronting inconvenient facets of liquidity provisioning in prediction markets is the second dragon platforms face; high-quality information has a price, and somebody has to be committed to paying that price for prediction markets to function as intended in thin settings. When we refuse to confront this dragon, and utilize microstructures poorly suited to the settings we deploy prediction markets in, our information channels become noisy (or entirely unreliable) and undermine the quality of the forecasts produced.

Resolution: Locked in Principle, Fickle in Practice

For prediction markets to elicit the trader behaviours they're designed around, everything must converge to resolution. Although damaging, a prediction market can survive thin liquidity. It can survive clunky UX. It can even survive a messy topic definition if the resolution is unambiguous and credibly enforced. However, although resilient, a prediction market cannot be expected to serve its purpose if resolution becomes discretionary after positions are established; such a market stops being about the topic being predicted and becomes about the system.

The biggest platforms today (e.g., Polymarket) resolve via some combination of human interpretation of rules, discretionary "clarifications", token-holder (or stakeholder) voting in an oracle game, and dispute processes that depend on who shows up.

These mechanisms can work in the sense that they usually produce an outcome, but that's an insufficient metric for success in resolution mechanisms being deployed in markets with real money, entangled incentives, and externalities. When designing an resolution mechanism for prediction markets, the objective was never simply "get an outcome". It is to get an outcome via a mechanism that is at a minimum: precommited upon market creation, reproducible a posteriori, verifiable by third parties, and resistant to capture when stakes are high.

Satisfying all of these objectives in tandem is hard, or arguably impossible, for most topics where resolution requires an interface with the outside world. Praxis tends to override the thesis here, and it's hard to give platforms a good finger wagging when you lack a strongly justifiable alternative. In turn, many platforms have embraced oracle designs that're good enough most of the time while satisfying additional properties that're desirable for themselves (e.g., "escape hatches" under sufficient blowback) and their users (e.g., fast resolution).

The end result is platforms relying on oracles that are more political than mechanical, more malleable than topics typically require, and designed under trust assumptions ill-suited (e.g., dependent on individuals' influence or patience) for the scales markets see today. Notably, these oracles are incompatible with the thesis the moment they allow discretionary change after meaningful open interest accumulates in a market since the platform has functionally introduced a new tradable asset—governance and interpretation risk.

A standard response to my critique might be "we need flexibility because natural language is messy". This is absolutely true...if you insist on building platforms that attempt to support "markets on any topic". Moreover, even if that is true, we should at least be honest about what we've built. We haven't built an oracle that makes truth legible on-chain, we've built an institution that interprets truth under incentives. In practice, regardless of how elaborately we dress them up in crypto-jargon about decentralization and trust, we've reintroduced a trusted intermediary in the moments it is most profitable to be one.

In markets where predicting the topic outcome is harder than predicting systemic behaviours, sophisticated traders focus on the latter; their strategy is a natural equilibrium response when resolution is not credibly locked. Therefore, to remain on a trajectory aligned with the thesis, we must design systems that aim to ensure that "predicting the system \(\approx\) predicting the topic outcome".

At this point it may seem like we've encountered an insurmountable dragon, but I disagree wholeheartedly. At a surface level it may sound like I've presented a critique inherent to these oracles; this is only true when we hold everything else fixed. As previously mentioned, prediction markets are tools and tools should be used where appropriate. The oracles used today require humans and those humans can participate in prediction markets, so incentives for earnestly performing your duty in the oracle can degrade dramatically when prediction markets reach a certain scale—platforms have indirectly incubated this dragon due to their own mainstream success, and they're now being forced to confront it. We're thus presented with three forks in the road: Recalibrate incentives (unrealistic and unsatisfactory), abandon the thesis (unsavory and socially damaging), or solve the problem (worry not, more on this in Part V).

Settlement: An Off-Ramp by Design, A Gatekeeper by Implementation

Even with well-defined resolution mechanisms having all the properties we can dream up, settlement can reintroduce trust on the way out. Platforms can satisfy the thesis from creation through resolution, and still rely on a final discretionary step for payout—a multisig, an operator transaction, an admin-controlled contract, a custodial ledger, or a centralized exchange mechanism that is ultimately enforced by internal databases rather than by on-chain settlement.

Why would this ever happen in practice? Intuitively this seems easy to avoid, and like a ridiculous obstacle to get tripped up on right at the end. The answer is painfully straightforward once we remember that modern systems exist in the real world while maintaining the illusion of having "markets on any topic": If your platform has taken on that role of a publisher and editor, then settlement is where operational and legal risk concentrates. (You were warned that dragon would keep rearing its head...)

When your platform gatekeeps market creation and encourages inappropriate or maliciously defined topics, then you better be able to delay settlement, freeze it, or apply exceptions to protect yourself when the unexpected (expectedly) happens. This becomes doubly the case when your goals as a "publisher" incentivizes you to push boundaries in pursuit of intrigue and spectacle.

From the perspective of the thesis, this is both entirely avoidable and absolutely fatal. If settlement requires permission, then these prediction markets are not mechanisms that settles by construction. It is a product that settles by permission, and whose payouts are contingent on an operator’s willingness (or ability) to execute them. The moment users internalize that possibility—even as a low-probability risk—it implicitly bleeds into the forecasting task. As we've repeated ad nauseum by now, a probability that includes the chance a platform will intervene is not a probability of the external event; it is a probability of some calibration about the event and of a platform-mediated payout event.

A Case Study: The Augur Post‑Mortem Fallacy

To help make these diagnoses more concrete, let's revisit a well-known early example in the history of prediction market platforms: Augur.

Thus far we've talked about the four control axes as if they're independent, but in reality they’re entangled. Open-ended market creation changes what "liquidity" means, topic ambiguity eventually spills into resolution, and slow resolution becomes a settlement problem. Augur exemplifies this coupling "nicely".

They tried to take the early thesis literally: Anyone could create a market on almost anything, markets traded on an order book, outcomes were resolved by an on-chain dispute game, and payouts were enforced by "code is law". For a moment, it may have even looked like the promised neutral substrate. However, its slow demise isn't from any one isolated dragon—it was these dragons reinforcing one another throughout the whole control surface.

We'll first consider the interaction between their choices surrounding market creation and liquidity. They went with "anyone can make a market" while also utilizing a CLOB as their microstructure, which results in a system that ensures anyone can fragment attention. When markets proliferate, the platform splinters into a long tail of thin books; in those settings a CLOB often has no meaningful price at all. The natural-but-wrong takeaway was "we need more volume". However, the more accurate takeaway is "order books need concentrated two‑sided flow", so a platform that encourages a long tail of user-defined markets needs (i) a microstructure/liquidity model that degrades gracefully in thin settings, and/or (ii) topic constraints that actually let attention concentrate.

Next, we'll hone in on the interaction between market creation and resolution. Augur allowed markets to be created using essentially unconstrained natural language, which gave them "expressiveness" in exchange for ambiguity, hard-to-detect duplicates, and adversarial markets. Not only did this hurt UIs built around Augur, the ambiguity resulted in a surge of costly disputes; resolution was often slow and socially exhausting.

Lastly, Augur's pipeline for market creation, resolution, and settlement created friction together. Slow resolution slows settlement down too, and similarly their gameable resolution mechanisms in turn made settlement gameable. Augur’s history includes "invalid market" dynamics where market creators exploited subtle inconsistencies to steer toward an "invalid" resolution, which changed payoffs in ways unprepared traders didn’t anticipate. So we saw that, even if specific components are nominally on-chain, the system can still produce outcomes that feel like rug pulls because the binding contract is the mechanism and not the plain English topic description.

In a sense, modern platforms seem to have arisen as a "minimum-change" response to Augur—they keep the familiar CLOB experience, concentrate liquidity by managing a catalog of high-engagement markets, and smooth over resolution/settlement with institutional processes that can move quickly under stress. The industry seems to have looked at Augur's mistakes and saw a set of single-axis lessons, but this led to incomplete conclusions; they fixed CLOBs with more volume, fixed the unconstrained topic problems with moderation, and fixed resolution speeds with discretion. By contrast, if you take a step back and look at Augur as a coupled system then different lessons fall out:

  • Microstructure needs to match topic scope. If you want a long tail of niche markets, then a pure CLOB is brittle without explicit liquidity commitments.
  • Guardrails are not synonymous with editorial discretion. Building appropriate topic guardrails is a protocol design problem, whereas curating topics is a business decision.
  • Speed can’t come from opaque malleability. If your resolution is too slow because of disputes, then speeding it up by making it malleable and discretionary after money concentrates isn't addressing the root cause.

Ultimately, building around incomplete conclusion clearly made the product work—but it also tightly wound the coupling until a new set of incentives formed. Which brings us to the next section: The gambling attractor isn’t a moral failure or a branding choice—it’s the equilibrium you drift toward when the easiest way to survive is to dance around the dragons rather than slay them.

The Gambling Attractor: Exchanging Truth-Seeking for Profits

Now that we've laid out the dragons that live enroute to building a prediction market platform, we're ready to say the quiet part out loud. When we embark on this journey, there are two clear attractors we can expect to reach:

  • A platform sustained by the informational utility it provides and the ideals it enshrines.
  • A platform sustained by the entertainment it provides and the profit it ensures for stakeholders.

The former corresponds with building towards the thesis, but is technically more difficult to implement and requires practically unbudging resolve. The latter corresponds with building while sailing around the dragons we've identified, and is unfortunately much easier to arrive at.

The point I've been making is not that any one choice that steps away from the thesis is inherently "bad". It's that each choice is a perfectly rational response to operating a platform as a business under real constraints. Once made, those rational responses gradually become entrenched as the platforms grows and infrastructure is built, which systematically moves platforms away from the promises they set out to fulfill. With time the economic incentives of the dominant platform model pushes prediction markets towards gambling even if everyone involved sincerely believes in the ideals laid out in the thesis.

In Conclusion...

The criticisms we've laid out are not an inevitability, they're a diagnosis we can systematically unravel and address. Building prediction market platforms without compromising ideals doesn't require philanthropy nor altruism, it just requires a strategy and innovation using modern tools that have only recently come into existence.

Some readers may wonder whether the concerns here are just pedantry, but it is important to remember that these influences compound and become self-reinforcing with the many factors we've discussed. Taken together, polluting the channel of information while eroding (or altogether replacing) the pillars our thesis is built atop is how we accidentally transmute forecasts into casino chips.

So what should we conclude? Well, in my opinion, it's that modern mainstream platforms are not slowly approaching the original on-chain thesis; they're optimizing toward a different equilibrium.

The stage is now set for the question we'll focus on for the rest of this series:
If the dominant branch of the prediction market platform ecosystem will keep sailing around these dragons—because it is rational to do so—what would it take to finally slay them?

Part IV:
Survival and a Need To Redefine Monetization

We need to address an elephant in the room. The previous part ended by describing the "gambling attractor"; a set of incentives pulling platforms toward markets that maximize engagement, volatility, and volume—even when it corrodes the platform’s ability to act as an epistemic utility. So, how do prediction market platforms make money? How can we avoid falling back onto a trajectory heading towards the "gambling attractor" because we need to keep the lights on?

The often unspoken tension we need to resolve hinges on a simple dichotomy:
Truth is the goal of prediction markets, but profit is the goal of platforms. When those goals conflict, the platform is who has the steering wheel.

How do existing platforms make profits?

Usually platforms today make money from some combination of transaction fees, spreads (directly or indirectly), liquidity provision advantages, and the compounding effects of higher volume and repeat engagement. None of these profit sources are inherently problematic, but none of them work to directly align a platform's goals with a prediction market's goal. In fact, quite the opposite, these alone work to cement an awkward relationship between truth-seeking and profits:

  • The more certain a market becomes, the less there is to trade.
    As probabilities converge, incremental edge is small, and trading activity decays. From the perspective of information aggregation, this is success. From the perspective of platform revenue, it's diminishing returns.
  • Ambiguity and controversy extend the trading window.
    Markets that are linguistically messy, socially contested, or procedurally dispute-prone generate long-lived volatility and, depending on the resolution mechanism, can keep the market open for longer. Volatility generates volume, and volume generates fees while fertilizing conditions for market makers.
  • Entertainment scales better than research.
    The public enjoys taking sides on salient questions, and emotionally-charged (e.g., divisive) topics are strong motivators for enticing participants who conflate their biases with an informational edge. By contrast, regardless of topic, far fewer participants do the due diligence needed to maximize information-rich trading. Platforms that need to maximize volatility and volume will rationally optimize (market catalogs, UX patterns, etc.) accordingly—even if that participation is not information-dense.

Under the current platform model, attention drives profits—not truth-seeking. If we keep adopting the same revenue levers, then we shouldn't be surprised when platforms slowly converge towards the same product; casinos wearing a prediction market-themed skin.

A solution begins revealing itself when we look at the difference between mental models:

  • The thesis calls for platforms as infrastructure supporting prediction markets that yield an informational public utility.
  • Popular platforms today behave like venues supporting prediction markets that yield an engagement-maximizing product.

Can we build a prediction market platform whose profit model looks like infrastructure? I think the answer is yes, but it requires designing the economics and the UX around a different center of gravity.

Profitability Without Betraying the Thesis

The crux of what we need to do is stop treating "platform profit" as something that must come from volatility, and start treating it as something that can come from usage of credible primitives. In other words, if the platform is truly infrastructure, then the platform should get paid accordingly—for execution, settlement, verification, etc.

As a good point of comparison in web3, we can look at something like Ethereum. Ethereum doesn’t need to curate "interesting transactions" to be profitable, it needs people to use the network; money comes from being a substrate for many workflows, not from editorializing which apps are allowed. A prediction market protocol can be similar: Profitable because it is a modular and credibly neutral substrate that people build on and rely on—not because it garners attention.

A natural objection is that infrastructure is only profitable if someone uses it—so who would actually use a credibly neutral prediction‑market network? I'd argue that demand for epistemic infrastructure is often latent until it becomes cheap and ubiquitous. New tools usually look silly or niche at first; then a few undeniable use cases break through; then the tool becomes a habit; eventually it becomes an expectation. Prediction markets are in the first part of that curve: They’ve penetrated the mainstream, and people have been forced (sometimes reluctantly) to notice that prices can compress information into legible forecasts in ways that punditry and polls haven't. The next step is making this utility available on demand—so these probabilities stop being spectacles people visit on a handful of headline markets and start being something they query whenever a decision needs to be made. The transition here will depend as much on public sentiment as it does on technical infrastructure; these tools can only succeed once people are willing to rely on them as legitimate instruments, and not dismiss them as gambling dressed in self-righteous mathematical jargon.

As we turn our gaze towards monetizing like infrastructure, we start to see revenue levers far more compatible with the thesis. There are numerous possibilities when workflow execution becomes our target, but here are a few notable options I wanted to highlight:

1) Protocol fees that scale with utility
A small and predictable protocol fee on trades, market creation, etc. is not inherently antithetical to truth-seeking—as long as it is credibly neutral. Protocol fees should behave like "boring" infrastructure tolls, and not as dynamic monetization tools. These fees need to be contractually fixed at market creation (not discretionary or modifiable to disallow impacting payoff structures), transparently structured, and non-discriminatory across topics and users.

Importantly, I'm not saying that prediction markets themselves can't have interesting fee structures. I am saying that platform fees and market fees inherently serve distinct purposes so should be treated as distinct entities. For the platform, fees exist to offset operational costs. For prediction markets, fees are mechanism design instruments that market creators can leverage to elicit desirable behaviours from traders. As a concrete example, for prediction markets in practice a flat fee can function like choosing numerical precision for market forecasts; if a flat fee of 0.01 is applied to every trade, then rational traders have no reason to buy shares at over 0.99 since it's a guaranteed loss.

2) Resolution and verification services (as commodities)
If the protocol supports multiple resolution mechanisms (fully on-chain deterministic, verifiable off-chain execution, refereed delegation, etc.), then resolution itself becomes a market. Users can choose the mechanism appropriate for the market they want to create, execution networks compete on price/latency, and the protocol takes a small cut as the pipeline coordinator. Under this model the platform no longer needs to act like "the house" and can focus on being "the plumbing". In addition, resolution mechanisms (aka oracles) can serve a purpose outside prediction markets, and these tools can be utilized broadly where appropriate.

3) Market creators pay the price of information
High-quality information has a price, and we should stop pretending otherwise. If a user wants a strong forecast, they should expect to pay for the liquidity (at least partially) and for overhead of all components required to elicit their forecast. Since the goal is to build a platform where all commitments are locked at market creation, we can make these costs explicit before a market ever goes live. The platform no longer needs volatility to stay alive, and earns revenue by providing a clean interface between capital and information.

Encouraging Passive Participation Without Becoming Casinos

Beyond re-imagining profit models, we also need to design appropriate avenues for participation. To be useful at scale while sticking to the thesis, we don't need every participant in the prediction market to be a professional trader or full-time researcher. Instead we need avenues for users to participate passively and reasonably—in ways that subsidize information aggregation without subsidizing degeneracy.

I've been thinking about several open questions related to how we can achieve this, and there are many exciting avenues to explore. Two promising directions (and doable in reasonable timeframes) are the following:

1) Constrained bidding languages to support participation
Imagine allowing users to allocate capital in a "vault" that can express a constrained family of actions, e.g. "In all markets of template T, buy up to X dollars of YES whenever implied probability is below my stated belief curve".

Building these tools turns participation into something akin to setting a policy. Users are still generating informative order flow, are able to express beliefs they expect to be true in expectation, and can do so passively. The key here is that strategies are expressed in a deliberately limited language; a language that disallows meta-strategies such as latency games and dependence on platform behaviours. If designed well, we can turn passive users into a form of epistemic liquidity rather than exit liquidity.

2) Liquidity pooling at market creation
By design, market creators often won't be interested in gambling since they are paying to try and learn something. As such, we should allow them (and others) to fund a market like you might fund a bounty; a market creator escrows liquidity (more on this in Part V) and allows others to "co-sponsor" the market by adding liquidity to the same pool, then the market becomes a public utility for a niche question—funded by the people who want the answer. Another way of looking at this is as a mechanism for empowering users who need to "crowdsource" information gathering. People already donate to open-source software and public goods, so why not allow them to donate to a forecast they rely on?

In Conclusion...

However much I wish it weren't so, righteousness and superior technical elegance aren't sufficient for keeping the lights on. As such, in we briefly touched on how prediction market platforms that build towards the thesis can survive and monetize. There's a lot more to say on this topic, and a lot to figure out. However, we know it's possible and have seen examples of success in related endeavors.

We'll likely come back to this topic several times in the future, but the main takeaway is that to succeed in the long-run we don't need builders that are some kind of saintly being who've sworn off material wealth. We simply need builders that are competent, determined, and innovative because the monetization and tokenomics are already workable. With that concern out of the way, we're ready to jump into the final part of this post talking about how we can build the prediction markets we were promised.

Part V:
Returning to the Thesis

I've tried being painfully explicit about a pattern I think many readers can feel but rarely name: Today’s prediction market platforms are converging toward designs that are more centralized, more permissioned, and more trust-based as they scale. This trend isn't because of some evil puppet master pulling strings, but because the path of least resistance is paved with discretion—discretion in market creation, discretion in liquidity provisioning, discretion in resolution, discretion in settlement. Unfortunately, although convenient for platform operators, discretion is exactly what prediction markets shouldn't allow once money arrives.

With the discussions thus far as a backdrop, outlooks about the feasibility of attaining the prediction markets we were promised might seem rather bleak. However, despite seemingly being adamant about the opposite, I truly am optimistic because I believe building the prediction markets we were promised has only become possible very recently and is worth doing.

In this part we pivot back towards optimism. We'll return to the question at the heart of this series:
How do we build platforms that give us an attainable path towards slaying these dragons and materializing the prediction markets we were promised?

As I see it, the missing pieces exist using some combination of a new generation of tools; chains that are cheap enough to run real workflows without hiding key components off-chain; verification techniques that make "commitment" enforceable rather than aspirational; and AI systems that can be embedded into oracles in ways that are both reproducible and verifiable. Furthermore, beyond this modernized technical toolkit, we are also at an inflection point of good will with the wider public—people are told prediction markets differ from traditional gambling, and are looking to understand their utility as an information source.

If brought together correctly, we have the ingredients needed to forge our own proverbial Balmung. Hopefully we can get it right without jumping the gun and souring public sentiment (as we've confronted in web3 before), as that's a necessary catalyst for reaching the adoption our decentralized technologies often require to serve their functionality and bring real value.

A Modern Dragon Slayer's Toolkit

At last, all our foundations are in place—we know what we need to build towards (Part II), what we definitely shouldn't do (Part III), and how we'll sustain ourselves when we finish building (Part IV). We're ready to talk about the tools we can lean on to slay the dragons getting in our way.

Constraining Market Topics to an Auditable Class of Claims

As we discussed at length, permissionless market creation without topic constraints is a trap. You either end up allowing everything (and invite markets that are unethical/illegal/coercive) until eventually getting crushed by regulatory pressure, or you curate markets manually (and reintroduce a central point of failure).

The way out of this trap is embarrassingly straightforward:
A permissionless prediction market platform must be restrictive about what is expressible. Not restrictive in who can create a market, but restrictive in what a market is allowed to be.

We can achieve this cleanly and scalably by defining a market specification language. Instead of letting market creators write arbitrary natural language (e.g., "Will X happen by Y?"), we define a constrained format for specifying markets via a flexible set of templates. Each template has a well defined outcome space, an allowed set of data sources, explicit time windows/timezones, appropriate resolution procedures to choose from, and clear-cut settlement rules. If something isn't expressible in the specification language, then it's not a valid market. Furthermore, with time, we can extend the specification language when underlying technical piping becomes able to support it, security surfaces are better understood, and users' needs become clearer.

Example of a Canonical Topic Class:
Gensyn testnet users may have engaged with Delphi, which demonstrates a fertile class of topics—ML model evaluations. We're immediately able to reduce ambiguity while allowing market creation to remain expressive. Market creators can express outcomes by choosing models from an approved list. Market tasks can be specified by choosing evaluation harnesses alongside scoring metrics, and choosing data sources is equivalent to committing to a dataset/benchmark/etc. Resolution mechanisms can be similarly defined via the type of oracle needed, how they're run, how a winner is chosen, etc.

Constraining market topics isn't about closing doors, it's what gives us a focal point to build around and can even open new possibilities by leveraging unique properties of specific topic classes. General prediction markets let traders present information through prices, but allowing traders to directly affect the outcome (especially with "markets on any topic") is typically considered a very bad property. However, when the topic class is chosen well, this no longer needs to be the case and we're able to utilize prediction markets as a form of crowdsourced search algorithm! Suddenly the space of information design opens up and markets can yield more than just a forecast. In the best case, this turns prediction markets into something closer to an open, incentivized research process.

Example with Prompt Engineering:
Suppose traders in a LLM evaluation market are allowed to submit prompts while buying shares, and they know that a model's max score across all submitted prompts is what'll be used for deciding the winner at resolution. The market now aggregates beliefs about which model will win AND incentivizes participants to search the space of possible prompts to maximize the performance of models they believe in. By making all submitted prompts public record, anybody with a similar task/benchmark can later come to this market for inspiration on how to best prompt engineer a model they care about for a task.

CFPMs and Ensuring the Price for Truth is Paid

A simple mantra summarizes the key takeaway of this subsection:
Always-on probabilities require always-on liquidity. Always-on liquidity requires committed capital.

In earlier discussions we talked about how CLOB-like microstructures allow liquidity providers to flee precisely when conditions aren't favorable to them, and this encapsulates why I believe CFPMs are the economic backbone needed for the thesis to work. Once a platform supports multitudes of markets being created on niche questions, a majority of markets will be thin—and thin markets need to facilitate trader(s) participation whenever a trader is interested (beggars can't be choosers after all) for the market to serve its purpose.

We've previously discussed how OMPMs (e.g., CLOBs) have an intrinsic failure mode for information aggregation—liquidity is voluntary and least reliable precisely when the market becomes most informative (near resolution, under controversy, during adverse selection). In flagship markets (e.g., big elections), where having sufficiently high-volume is a reasonable assumption, OMPMs are arguably more appropriate even if some trading becomes entangled with microstructure meta-strategies. However, in thin market settings, these failure modes are all but guaranteed to show up and can outright kill thin markets by discouraging potential traders from ever participating.

CFPMs are designed with the antidote in mind; liquidity is a market parameter guaranteed via an escrowed commitment. When a market is created, someone escrows collateral and chooses a liquidity parameter (e.g., \(b\) in LMSR). In turn, the committed liquidity ensures two critical properties:

  • Prices exist at all times. The probability vector is continuously defined throughout the entire market's duration.
  • The information is already paid for. No counter-parties are needed, and market creators have rigid bounds on their worst-case loss.

Interestingly, CFPMs naturally give market creators a clean interface for controlling the accuracy of their markets via the amount of liquidity they commit, i.e. more liquidity allows markets to be more reliable and informative. From this perspective, you see that my statements about "paying the price for truth" weren't just flowery language or metaphorical hand-waving. By making market creators commit liquidity at market creation, our information aggregation tools become an explicitly priced resource.

Before moving on I wanted to touch on a common question I expect to see at this point: "Doesn't the market maker lose money?" Sometimes, yes. That's kind of the point. In CFPMs the liquidity provider is not a passive rent-seeker harvesting spreads. Instead, its more appropriate to think of them as underwriting the creation of a collective forecast—they're paying a bounded loss in exchange for making the market exist. Coincidentally, this distinction is also a differentiator in how public utilities get funded compared to how casinos get funded.

Reproducible and Verifiable Resolution Mechanisms

I’ve talked at length about how discretionary resolution pollutes everything upstream with governance speculation. The understated part is that discretion is most prescient during edge cases—when the money is large, wording is strained, the world is messy, and incentives are adversarial. Historically, this is where purely algorithmic approaches tend to crumble and it's where platforms justify discretion. Modern problems call for modern solutions, and luckily those solutions have become available to us very recently.

The goal here is easy to state and difficult to achieve:
Design resolution mechanisms that are robust to edge cases and can be replayed and verified. Then, at market creation, commit to one of these mechanisms so that discretion is removed from the pipeline.

To make this goal attainable, we need to change our mental model. Oracles are not institutions we rely on; they’re programs we compose. Under this view, a resolution mechanism is best understood as a state machine orchestrating the flow of inputs and outputs through a set of pre-committed "oracle programs". These oracle programs can take on many forms—traditional scripts, ML models, etc.—but it’s convenient to treat them uniformly as black boxes whose inputs are pre-committed evidence bundles (data sources, hashes, timestamps, etc.) and whose outputs are a market outcome (or signals the state machine uses to advance, escalate, or invalidate).

Despite how we often want to talk about them, a resolution mechanism’s job is not to be "smart" nor "correct". Their most important role is actually to be binding:
To ensure that, for each market, there exists a fully specified, tamper-resistant procedure that (i) consumes committed inputs, (ii) deterministically selects which oracle programs to run and when, and (iii) produces a final outcome.

For the purposes of this post—and because I believe they cover a vast majority of use cases—I’ll assume oracle programs fall into three broad classes:

  1. Deterministic programs, e.g. scripts over structured inputs running "simple" algorithms.
  2. ML oracle models, e.g. an LLM judge that maps evidence to a decision.
  3. Multi-oracle aggregators, e.g. voting/consensus rules or Schelling-like mechanisms over multiple oracles' outputs.

With just these building blocks, resolution mechanisms become robust procedures we pre-commit and compose from well-designed parts. For example, a resolution mechanism might default to a cheap deterministic check, then escalate to an LLM judge when the case is linguistically messy, and finally escalate to a panel of LLM judges (with a pre-committed aggregation rule) when we need redundancy against individual models' failure modes. Crucially, this can all happen without human discretion; every escalation path is specified up front, every oracle is pre-committed at market creation, and every edge case is handled by a predefined state transition.

Circling back to our goal for this section, we need to argue these resolution mechanisms can answer two key questions.

In what sense are we robust to edge cases?
First, it's important to clarify that, when I say "robust to edge cases", I do not mean "the oracle will always find the metaphysically correct answer". In practice, some questions are genuinely contested, some topics are inherently subjective, some evidence is noisy, and some events don’t map cleanly to any discrete outcome set. So, what I mean is much narrower and relevant to prediction markets; I consider a resolution mechanism robust if it remains well-defined when the world is ill-defined.

Intuitively, when ambiguity emerges, a robust resolution mechanism executes a pre-committed procedure without intervention and—importantly—has an explicit way to end. Ensuring resolution mechanisms reach a reasonable terminal state within an acceptable time span is one of the central motivations for why platforms bake discretion into their resolution mechanisms, and even with discretionary resolution they still require "no-op" outcomes to keep resolution timely. If a market can be argued about indefinitely, then resolution reduces to a game of politicking and patience. However, avoiding this by allowing "clarifications" ex post, brings us back to governance speculation contaminating the market.

We can sidestep these issues by letting resolution mechanisms define a ladder of escalations, which pre-commit the mechanism to a sequence of increasingly powerful tools with clearly defined failure modes and termination conditions. We could, in principle, design mechanisms that jump straight to the final escalation step and work, but in most circumstances such an approach is overkill and needlessly costly.

Different oracle programs have different strengths, and you want to exploit those strengths rather than strapping the market’s integrity on any single approach:

  • Deterministic programs are cheap, legible, and easy to verify. They excel when the topic is already structured and the evidence is formal (on-chain events, signed feeds, benchmark scores, etc.). That said, they're brittle under natural-language ambiguity, missing context, etc.
  • ML oracle models are expensive and less interpretable, but radically more flexible. They shine where scripts fail—messy language, context-sensitive interpretation, evidence weighing, and fuzzy categories. Interestingly, these settings are exactly where resolution mechanisms historically break in practice.
  • Multi-oracle aggregators sacrifice simplicity to gain reliability. A single oracle can be biased, misaligned, too rigid, too eager to decide, or just wrong in a correlated way. However, a multi-agent mechanism can dramatically reduce that risk by (i) introducing diversity, (ii) requiring consensus thresholds, and (iii) making the procedure more stable to any individual oracle's quirks,

Robust resolution mechanisms that work well in the real world likely won't hinge on just one oracle—they'll work well because they're built on a composition of oracles. A state machine that can escalate allows the resolution mechanism to do what a single oracle can’t. It allows mechanisms to absorb and redirect edge cases without needing human discretion to interpret rules ex post, and it allows for a structured "dispute" step that is still non-discretionary.

Toy example of what a generic resolution mechanism "program" could look like. When some pre-committed resolution condition triggers (time, proxy signal threshold, etc.) the mechanism starts at state 0 and can escalate until it either returns a valid outcome or seemingly can't. Escalations should be triggered by automated criteria OR something like a stake-backed dispute procedure (so the disputer, at a minimum, foots the bill for inappropriate escalations).

How do we guarantee resolution can be replayed and verified without reintroducing trust?
Now we get to the sharper question and, unsurprising if you're familiar with Gensyn, one we've thought about at length. Once we pre-commit to a resolution mechanism, how do we ensure third parties can replay oracle programs and verify their outcome?

This breaks into two inter-related subproblems:

  • Determinism (or, more generally, reproducibility) \(\rightarrow\) Can independent parties run "the same" oracle program and get the same output?
  • Verification \(\rightarrow\) Can an on-chain process ensure the reported output is correct without trusting any single entity?

Tractable (and satisfying) approaches to these problems have only recently become available in both classic sequential computation and ML inference. These days, the question isn't whether we can, it's about when we need to solve these problems and how strong of guarantees we require. I'll keep the discussion here very surface-level to keep ourselves on-track, but in a future blog post we'll tackle this topic comprehensively.

Enforcing determinism in "normal" programs is a well-understood engineering constraint, and there are numerous approaches to building around it. For ML inference there are fewer options and the space is evolving rapidly, but we have a spectrum of approaches from the basic tooling in Torch all the way to Gensyn's own RepOps library.

Now, even with determinism, somebody still needs to confirm the output of the oracle program. To avoid trusting a single executor or needing to run expensive computations on-chain, we can verify an oracle program's output using redundancy (e.g., re-run the whole program) or cryptographic protocols (e.g., zkML or Verde's bisection game).

At the end of the day, practical resolution mechanisms will likely take a hybrid approach where weaker guarantees are used by default and gradually strengthened under disputes. Such an approach allows us to run something cheap most of the time while, when needed, remaining robust to edge cases and flexible to disputes.

From Creation to Settlement: Committing to a Single On-Chain Lifecycle

A constant throughout this post is that we should aim to remove discretion at every stage of a market's lifecycle—from creation through settlement. The last part of our toolkit is arguably the least glamorous yet most important: Market lifecycle engineering.

The essence is to make market creation an intuitive and programmable primitive, one that pre-commits everything from the market topic through to settlement. Personally, I envision market creation as a three step process where the creator: (i) configures a market contract, (ii) commits to it's terms, and then (iii) passes it to a "market factory" that deploys an immutable live market.

In the first step, when a contract is configured, the creator must specify the topic template, the liquidity mechanism, the resolution mechanism, and so forth. Once configured the market creator will have an itemized cost (or at least an upper bound) for starting their market, and if they agree they must commit by escrowing funds accordingly. Finally, with commitments ready and a well-defined market, the creator deploys the market.

Once all three market creation steps are taken, and the market is live, the market should functionally be out of its creator's hands and run a predictable coarse:
Trading goes live \(\rightarrow\) Resolution mechanisms trigger, an outcome is determined \(\rightarrow\) The market contract accepts the outcome if the pre-committed verification rules are satisfied \(\rightarrow\) Settlement is automatically triggered and enables redemption/payout.

Ten years ago, folding the entire market lifecycle into a pre-committed contract would've been a herculean feat. Today, although certainly a tricky technical puzzle to assemble, web3 infrastructure and tooling has progressed significantly. Modern L2s, stronger contract engineering disciplines, etc. make it reasonable and scalable for us to manage every part of the lifecycle on-chain. Not everything has to be executed on-chain, but the chain is what drives markets along while ensuring commitments are enforced and states are transparent.

Closing Remarks

Writing this post wasn't my original goal; I simply set out to understand the current state of prediction market platforms and how they might set us on a path towards the markets I assumed we were working towards. While exploring, I quickly realized they're on a one-way trip towards centralization. After thinking about the forces pulling platforms in that direction, I came to a conclusion: Prediction market platforms haven't failed to build towards the thesis because it's infeasible to achieve. They've failed because platforms have made decisions that ensure strong commitments are existential challenges.

My criticisms are not about any one platform, and I'm conscious that platforms today took on the shape we know due to historical precedence and myopic decision-making under complex pragmatic pressures. However, when we look at prediction market platforms from first principles, we can envision a future leading towards the prediction markets we were promised.

Redefining the existing platform model, utilizing state of the art tools, and innovating to bring it all together will be challenging. However, it certainly is possible and the implications of building the prediction markets we were promised are vast. Slaying dragons isn't easy work, but I believe it's worthwhile and I know others do too.