The Index Problem

To classify is human. Not to classify is to be dead to the world. But the classifications we build are never innocent.

Geoffrey Bowker and Susan Leigh Star, Sorting Things Out, 1999


The tree of proof described in the previous chapter solves the problem of the gatekeeper at the substrate. It does not, by itself, solve the problem of the gatekeeper at the view. A reader on a phone cannot materialize millions of branches locally. Between every reader and the substrate sits a system that says here is what is on the tree, and here is what is worth your attention first. And that system is where the book’s argument runs into its sharpest remaining question.

Each architectural turn the book has named so far, the payment rail, the attention layer, the institutional record, has had a bottleneck reappear one layer up. The ratchet has one more turn in it, and the turn is the one everything above rests on.

In April 2026, Vitalik Buterin posted a warning on X. I saw it the morning it went out.

The kind people at @eth_limo have warned me that there has been an attack on their DNS registrar. So please do not visit vitalik.eth.limo or other eth.limo pages until they confirm that things are back to normal.

He directed his readers to his blog through IPFS. Bypassing the compromised layer entirely, pointing to the content rather than to a name that could be redirected.

A human reading that tweet could act on it. The warning was intelligible. The risk was legible. The workaround was a click. The content survived because it lived on a substrate beneath the attack.

I read it twice. Then I tried to picture what an agent would have done.

An agent that had been told to read Vitalik’s blog would have queried DNS, received the redirected address, and navigated to whatever the attacker had placed there. No warning. No hesitation. No judgment. The agent cannot read a tweet telling it the index has been compromised. It reads what the index serves.

The attack happened. The warning was issued. The human workaround worked. The agent workaround did not exist.

That gap, between what a human can navigate around and what an agent must read through, is the index problem. It is running now, against a population of agents that is growing faster than the discovery layer is being made safe.

The brand is the bottleneck. Not the protocol. The protocol is open: anyone can run a search engine or a DNS resolver. The bottleneck is the layer where reputation lives. Outside the surface the agent reads. Compromise the brand through registrar attack, algorithm change, or regulator pressure, and every agent reading the surface is misdirected at once. The protocol does not carry the trust. The brand does.

The Swarm Problem

A single misdirected agent is a manageable error. A misdirected swarm is a systemic event.

The agent economy will not be a collection of isolated agents making independent decisions. It will be swarms. Thousands or millions of agents querying the same discovery surfaces, importing trust from the same brands, acting on the same instructions. The coordination that makes a swarm powerful is the same property that makes it catastrophically vulnerable to a single point of compromise. Every agent that imports trust from the same place is misdirected from the same place.

This does not require a sophisticated adversary. The eth.limo attack was a registrar compromise. Not a novel technique, not an advanced persistent threat, not a state-level operation. It is the kind of attack that happens routinely against human-facing infrastructure and gets caught because humans can read warnings. The agent has no Vitalik to follow.

It does not require an external attacker at all. A ranking algorithm changes without notice, without disclosure, without any channel through which an agent can detect that the surface it was reading last week is not the surface it is reading today. From the agent’s perspective, an algorithm change that systematically surfaces certain results over others is indistinguishable from a deliberate attack. The brand says the same thing in both cases.

A misdirected swarm does not fail gracefully. The speed that makes agent swarms economically valuable is the same speed that makes a compromised discovery layer catastrophic.

The Bottleneck Migrates to the Index

None of this is new. HTTP is open and anyone can run a server, but a single search index has been the web for any reader who does not type a URL directly. DNS is distributed, yet registrars are centralized chokepoints. As the eth.limo attack demonstrated in a single afternoon. IPFS content lives on a peer-to-peer network, and most readers reach it through a handful of HTTP gateways. Bitcoin L1 is decentralized at the level any engineer can verify, and most queries about Bitcoin state route through a small number of block explorers. The substrate is distributed. The common view is not.

LLMs add one more layer of compression. The retrieval stack, what a model pulls for a given question at inference time, is the new index, operated by a handful of labs, with no reader seeing behind it. For the growing share of readers whose primary research surface is a model, the retrieval layer is the index of the index of the web. One more brand to take on faith.

The pattern does not require a conspiracy. Indexing has economies of scale, and the economies produce one dominant player per category. The protocol stays open. The brand consolidates. Whoever owns the brand owns the resource, because no one reaches the resource without taking the brand on faith.

Category Branches

The architectural answer is not to build a better brand. It is to move reputation onto the substrate.

In the structure described in the previous chapter, branches are identities. A weather oracle and a medical oracle hang off the same trunk with no native structure connecting them by category. Navigation between them requires an external indexer. Something that says here is what exists, here is what is worth your attention, here is where to look. That external indexer is a brand. It is the registrar of the tree.

A category branch is a different kind of branch. A substrate-level place rather than a brand-level pointer.

The shape is older than the internet. IP addresses deliver packets; street names tell you how to get somewhere. The internet has IP and never quite got streets. DNS gives names that resolve to addresses, but nobody owns the navigable hierarchy of place that streets give a city. Each platform built its own private streets and called them an index. The Tree of Proof is the missing public layer. A category branch is anchored at cost, maintained over time, weighted by the four forces. Its reputation is no longer a brand. It is a property of the street itself, readable by anyone walking the substrate.

Maps still get made. There will always be cartographers, Google, Apple, OpenStreetMap, the artisan publishing a coffee-table atlas, and a reader navigating the substrate is reading some kind of map. But no cartographer redraws the streets. The substrate decides where the streets are; the map-maker decides which to highlight. That is what reducing the power of the map-maker looks like in practice.

Time is on the chain. Cost is on the chain. Proximity is on the chain. Hash validity is on the chain. The agent does not have to import trust in the branch from outside. It derives the branch’s reputation from the same substrate it is already reading. The blindness ends. Not because someone gave the agent sight, but because the surface became legible at the level of physics. The indexer that points the agent to a branch is a pointer, not a judge. The chain is the judge.

This is also where competition reappears. A search engine has one ranking function, opaque and unilateral. A category branch can host many routers competing for weight, and the competition is auditable: an agent comparing two routers on the same branch sees their elapsed time, their cost, their hash validity. Brand competition is opaque. Physics competition is readable.

The Router Model

Category branches create a business model the existing index cannot. A review platform maintains a local-business branch. A medical consortium maintains a healthcare branch. A legal database maintains case law. A financial-data provider maintains market data. Each is dominant within its category and irrelevant outside it. The router’s reputation is the branch’s health. Time alive, cost spent, hash validity intact. A business listed there inherits the thermodynamic weight.

The revenue model inverts. Today a review platform pays a search company for traffic and charges businesses for visibility within a system the search company controls. On the tree, the platform is the discovery layer for its category. The search company cannot demote it because the search company does not own the substrate. The router’s continued dominance depends on maintaining the branch. Which means the economic incentive and the integrity of the discovery layer are aligned rather than opposed. If the router disappears, the business’s history on the branch survives. That is a different relationship between platform and business than anything that exists today.

What This Does to Google

The architecture that follows from category branches and routers does not compete with Google. It routes around it. The same way Bitcoin routes around the correspondent banking system without attacking it.

Google’s value to human searchers is relevance. Their value to agents would have to be verifiability. Provable completeness, provable freshness, provable integrity. They cannot offer this because the index is proprietary, centrally controlled, and unverifiable by design. The architecture that makes them powerful against human searchers is the architecture that makes them structurally inadequate for agents. The brand is the moat. The brand is the bottleneck. Those are the same sentence read from two sides.

The default-search slot loses its weight too. On the tree, the search box becomes a reader of a public substrate. Any company can build one. The data is open. The competition is in the reading. An agent compares two readers running against the same substrate and sees exactly why they returned different results. And selects the one whose weighting it can verify rather than the one whose brand it has to trust.

The yellow pages were not attacked. They were made optional by something that did the job more verifiably. By the time they understood what was happening, the migration was complete. Google will not be attacked either. Agent traffic will migrate toward discovery surfaces that can prove what they serve. Quietly, showing up as a number declining in a dashboard with no single event to explain it.

What This Does Not Solve

The chapter has to close on what the architecture does not do, or the claim will not survive the reader who is paying attention.

It does not eliminate concentration. Routers will still dominate their categories under market pressure. What changes is what dominance means. A dominant router that cannot lie about completeness, cannot lie about freshness, cannot redirect without leaving a thermodynamic trace, and cannot hide its signing history is a different kind of dominant than a registrar or a search index. Not a smaller shadow. A different shadow.

It does not eliminate cold search at the timescale this book is being published in. A reader looking for a branch they do not yet know exists still needs a discovery surface, and the surface will be provided by whoever has scale. I have sat with this the longest. Most of the people who will ever read what the tree carries will never run a node. They will reach for the surface the quickest path gives them, the way I reach for an explorer when I want to verify an inscription. The architecture has to be honest about that.

The bad-data-within-a-branch problem, a router can maintain a branch honestly and still list stale leaves, is real. It is taken up under its own heading in Seven Seams.

The bootstrap problem, that early actors shape the structure, is self-resolving. The chain does not read motive. It reads cost, time, proximity, and hash validity. The trunk thickens through use, not faith.

What it does is move reputation. The eth.limo attack worked because the brand was the trust, and the brand could be redirected by compromising the registrar that issued it. A category branch on the tree carries no equivalent brand. Compromising the discovery layer requires compromising the substrate. Compromising the substrate requires forging blocks. Forging blocks requires energy the attacker cannot fake.


Vitalik pointed to his IPFS address because the content existed on a substrate beneath the attack. The tree is that substrate, extended to the discovery layer itself. Not just for content. For the index that tells agents where to look.

The index can concentrate. The brand can be compromised. The brand can be quietly redirected by an algorithm change or a registrar attack with no warning channel that agents can read.

The answer is not a better index. It is a discovery layer where reputation lives on the surface the agent is already reading. Routers whose weight is enforced by physics. Category branches whose history is on the chain. The four forces, in place of the brand.

The agent that read the redirected eth.limo address did not know it had been redirected.

The agent that reads a category branch on the tree knows. Because the reputation lives on the surface the agent is reading, and the surface does not require trusting anyone.