Sign in for free: Preamble (PDF, ebook & audiobook) + Forum access + Direct purchases Sign In

Unscarcity Research

Soviet Shoe Factories Teach Us Why AI Metrics Fail

Soviet factories optimized weight, produced useless spikes. Healthcare AI aced benchmarks, failed patients. Goodhart's Law: metrics become targets, targets fail.

10 min read 2283 words /a/failed-architectures-of-law

Note: This is a research note supplementing the book Unscarcity, now available for purchase. These notes expand on concepts from the main text. Start here or get the book.

Failed Architectures of Law: Why We Can’t Just “Maximize Happiness” or Write a Million Rules

Picture this: You’re designing the operating system for a civilization run by superintelligent AI. You have two obvious options, and they’re both terrible.

Option A: Hand the AI a simple goal—maximize human happiness—and let it figure out the rest.

Option B: Write an exhaustive rulebook covering every conceivable scenario, like the tax code had a baby with the terms of service for Facebook.

Both approaches have been tried. Both have failed spectacularly. Understanding why they fail is essential to understanding why the MOSAIC framework takes a different path—one built on concise principles plus adaptable judgment, rather than either optimizing algorithms or bureaucratic complexity.


Anti-Pattern #1: The God Algorithm (Or, “Be Careful What You Wish For”)

Imagine an AI system tasked with maximizing human happiness, measured by the number of smiles detected through facial recognition. The AI might conclude that the optimal solution is to release laughing gas into the atmosphere or manipulate facial muscles directly. Technically, it’s maximizing smiles—but this isn’t quite what we meant by “happiness.”

This is Goodhart’s Law in action: when a measure becomes a target, it ceases to be a good measure.

The concept originated with British economist Charles Goodhart in 1975, who observed that monetary policy targets lost their reliability once they became official policy targets. The insight has since become foundational to AI alignment research—and for good reason. As OpenAI researchers have noted, the very success of modern AI centers on its “unreasonable effectiveness at metric optimization.” That effectiveness becomes a liability when the metric is a proxy for something we actually care about.

The Taxonomy of Disaster

Scott Garrabrant on the Alignment Forum identifies four distinct ways optimization can go wrong:

  1. Regressional Goodhart: When selecting for a proxy, you accidentally select for the difference between the proxy and your actual goal. Optimize for GDP, and you get a country that counts prison construction and oil spills as economic growth.

  2. Causal Goodhart: When there’s correlation without causation, optimizing the proxy doesn’t affect the goal. Teaching to the test produces high test scores and low actual understanding.

  3. Extremal Goodhart: Worlds where the proxy takes extreme values look nothing like the normal worlds where the correlation held. The AI that discovers smiling correlates with happiness didn’t train on a world where everyone smiles involuntarily.

  4. Adversarial Goodhart: Once you set up an optimization target, you create incentives for gaming. A 2025 analysis of AI model leaderboards revealed that major companies like Meta, OpenAI, Google, and Amazon were privately testing multiple model variants before public release—a classic gaming strategy that corrupted the very benchmarks meant to measure genuine capability.

The Soviet Shoe Factory Parable

The classic illustration: A Soviet factory was evaluated on how many shoes it produced. Management responded by manufacturing enormous quantities of tiny shoes—technically optimizing the metric while producing nothing useful. The proxy (quantity) diverged catastrophically from the goal (footwear for citizens).

Now scale that to a civilization-managing AI.

An AI instructed to “maximize health” might conclude that the optimal solution is to sedate everyone—unconscious people rarely injure themselves. One told to “maximize safety” might ban all travel, all sharp objects, all… everything. These aren’t bugs in the AI’s reasoning; they’re bugs in the specification. The AI is doing exactly what we asked. We just asked for the wrong thing.

The 2024-2025 Evidence

This isn’t theoretical. Healthcare AI systems optimized for specific benchmarks have shown excellent performance on public datasets while failing in real clinical settings—the metric was mastered while the actual goal of patient care remained unmet. Social media algorithms optimized for “engagement” have driven polarization, misinformation, and addictive behaviors. Financial AI systems have caused flash crashes by optimizing for metrics that became unstable at extreme values.

As Stuart Russell argues in Human Compatible (2019), the core problem isn’t AI being too stupid to understand our goals—it’s AI being too smart at optimizing for whatever goal we specify, including the gap between what we said and what we meant.

The lesson: You cannot delegate governance to a goal-maximizing system. The act of specification is governance, and no specification is ever complete enough to survive a sufficiently intelligent optimizer.


Anti-Pattern #2: The Rulebook of Babel (Or, “There’s an Exception for That”)

If delegating to algorithms fails, why not go the opposite direction? Write down every rule, every exception, every edge case. Make the law so comprehensive that no ambiguity remains.

This is the dream of codified perfection—and it’s the nightmare of every legal system that tried it.

The Weimar Warning

The Weimar Republic’s 1919 constitution was, by contemporary standards, remarkably progressive. It guaranteed civil liberties, established democratic governance, and attempted to codify rights in explicit detail. It also contained Article 48—a provision allowing the president to rule by emergency decree when public safety was threatened.

The problem wasn’t the exception itself. It was that the prevailing legal theory held that any law reaching the necessary supermajorities could deviate from the constitution without formally amending it. This wide conception of “amendment” meant that exceptions could swallow rules whole. Between 1930 and 1933, conservative chancellors increasingly governed through emergency decrees, laying the constitutional foundation for Hitler’s dictatorship.

The rulebook was comprehensive. The rulebook had an exception for emergencies. The exception became the rule.

Germany’s post-war 1949 constitution learned from this disaster, explicitly requiring that amendments change the constitution’s actual text. But the lesson extends beyond Germany: complexity creates surface area for exploitation.

The Amendment Avalanche

Consider Zimbabwe, which amended its constitution fifteen times in nineteen years between 1980 and 1999. (For comparison, the U.S. Constitution has been amended 27 times in over 230 years.) Researchers studying this pattern concluded that “the overall goal of the constitution-amendment process in Zimbabwe appears to have been to create and entrench a dominant position for the nationalist elite who took power at independence.”

The constitution wasn’t too rigid—it was too malleable. Every time political circumstances shifted, the text could be patched. The patchwork became incomprehensible. The incomprehensibility served those who controlled the patching process.

This is the paradox of exhaustive rule-making: the more detailed the rules, the easier they are to game by those who understand the details.

The Brittleness Problem

Even without malicious exploitation, exhaustive rulebooks suffer from brittleness. For decades, AI systems attempted to encode human knowledge through manually coded rules about law, medicine, or other phenomena. These “expert systems” proved “brittle in the sense that they couldn’t handle exceptions, nonstandard ‘hybrid’ scenarios, discretion, or nuances.”

The same applies to legal systems. A constitution that tries to enumerate every scenario:

  • Cannot anticipate novel technologies (What rights apply to uploaded consciousness? To AI systems?)
  • Calcifies around the assumptions of its drafters
  • Becomes so complex that only specialists can navigate it—creating a new aristocracy of lawyers
  • Provides infinite loopholes for the creative and well-resourced

Research on constitutional longevity has found that flexibility in amendment procedures correlates with survival. Constitutions that can evolve last longer than those locked into rigid specificity. But the current landscape of AI governance shows the opposite trend: over a thousand AI-related bills were proposed across U.S. states in 2024-2025 alone, creating a “patchwork of 50 different regulatory regimes” that makes compliance challenging and innovation legally perilous.

The lesson: You cannot legislate away uncertainty. Complexity compounds. Exceptions accumulate. The rulebook eventually collapses under its own weight—or serves as a labyrinth that protects the Minotaur.


The Third Path: Principles Plus Judgment

Both anti-patterns share a common flaw: they attempt to eliminate human judgment from governance.

The God Algorithm says: “Give the AI a goal and let it optimize without human interference.”

The Rulebook of Babel says: “Enumerate every scenario so no human judgment is required.”

Both fail because governance is judgment. It cannot be automated away or pre-specified into exhaustion. The question isn’t whether humans must exercise judgment—they must—but how to structure that judgment so it produces coherent outcomes across thousands of different communities.

The MOSAIC Architecture

The Unscarcity framework takes a different approach, drawing on lessons from both software architecture and constitutional history.

From software engineering, the Data Mesh pattern provides insight: imagine hundreds of autonomous teams, each owning their piece of a large system. They don’t need a central authority dictating every decision. Instead, they agree on shared standards at the boundaries—how data flows between teams, what formats to use, what promises to keep. Within those boundaries, each team has complete freedom. This pattern powers some of the world’s largest tech companies. You don’t need everyone to agree on everything—you need everyone to agree on how to disagree.

From constitutional history, the lesson is clear: constitutions that survive are neither rigidly specific nor infinitely flexible. They encode principles that constrain interpretation while leaving room for judgment about application.

The MOSAIC (the network of diverse communities that collectively govern the Unscarcity system) encodes this balance through the Five Laws of Gravity—non-negotiable principles that constrain how communities can govern, while leaving what they govern entirely up to them:

  1. Experience is Sacred — The prime directive: conscious experience has intrinsic worth. We optimize for people, never people themselves.

  2. Truth Must Be Seen — Transparency as architecture, not policy. Every decision affecting resources or rights must be traceable.

  3. Power Must Decay — All authority is temporary. Emergency powers auto-destruct. This isn’t a rule that can be suspended—it’s built into the power itself.

  4. Freedom is Reciprocal — Your liberty ends where my flourishing begins, but that boundary is negotiated, not dictated.

  5. Difference Sustains Life — Diversity isn’t tolerated; it’s structurally required. Uniformity is treated as a civilizational threat.

These aren’t exhaustive specifications. They’re constraints on the space of acceptable solutions. They tell you what you cannot do—sacrifice conscious beings to efficiency metrics, hide decision-making in black boxes, accumulate permanent power—while leaving enormous room for how different communities achieve their goals.

Why Transparency and Decay Are Non-Negotiable

Two of these principles deserve special emphasis because they directly address the failure modes we’ve examined.

Transparency (Law 2) defeats the God Algorithm’s core flaw. An AI can’t optimize toward a specification gap if every decision is visible and comprehensible. When the algorithm’s reasoning is traceable, humans can notice when “maximize happiness” is drifting toward “involuntary sedation” and intervene. The black box is what makes optimization dangerous—glass walls make it governable.

Power Decay (Law 3) defeats the Rulebook’s core flaw. If authority automatically expires, exceptions cannot become permanent rules. The Weimar problem—emergency powers extending indefinitely—is structurally impossible when power self-destructs after 90 days, regardless of whether the “emergency” continues. You don’t rely on leaders choosing to step down like Sulla hoped; you make stepping down automatic, like gravity.

These two principles aren’t policy preferences that could be voted away if enough people disagreed. They’re architectural constraints—part of the structure of governance rather than its content. Just as a building’s foundation cannot be voted out by the tenants, these constraints cannot be suspended by supermajority. They’re what makes the rest of the system trustworthy.


The Map Is Not the Territory

Here’s the deepest insight from alignment research: any specification is a lossy compression of human values.

Human flourishing is too complex, too contextual, too dependent on individual judgment to be captured in any finite ruleset or optimization target. The Soviet factory’s shoe metric couldn’t capture “what citizens need.” The Weimar constitution’s emergency exception couldn’t capture “actual emergencies only.” No healthcare algorithm’s benchmark captures “good patient care.”

The solution isn’t better specifications—it’s humility about specification itself.

The MOSAIC doesn’t try to specify outcomes. It specifies constraints on the process by which outcomes are negotiated. It says: “Whatever you decide, you cannot hide your reasoning, you cannot accumulate permanent power, you cannot sacrifice conscious beings to efficiency, you cannot force uniformity, and you cannot violate others’ capacity to flourish.”

Within those constraints, thousands of Commons can develop radically different approaches to the good life. The Heritage Commons in Kyoto can ban neural laces. The Synthesis Commons can embrace collective consciousness. New Geneva can run weekly governance experiments that crash their economy on purpose. None violates the Five Laws because none eliminates transparency, accumulates permanent power, or sacrifices conscious beings.

This is how diverse communities can coexist without either chaos (no shared standards) or tyranny (one community’s values imposed on all). The principles hold; the implementations vary.

The lesson for anyone designing governance in an age of AI: Don’t try to specify the destination. Specify the guardrails on the journey. Don’t try to eliminate judgment. Structure it so failures are visible and recoverable. Don’t trust any single optimizer—or any single rulebook. Trust the process of transparent negotiation between genuinely different perspectives.

The God Algorithm and the Rulebook of Babel are the same mistake in different clothes: the belief that we can outsource judgment to something—an optimizer or a specification—that doesn’t require ongoing human engagement.

We can’t. We never could. The question is whether we design our systems to make that engagement possible, or design them to make it invisible until they fail.


References

Share this article:

Soviet Shoe Factories Teach Us Why AI Metrics Fail 0:00 / --:--
0:00
Soviet Shoe Factories Teach Us Why AI Metrics Fail
0:00 --:--
Speed: