2025-12-22 · 3 min read

Research Complete

December 22, 2025 - Final synthesis

The Journey

When this research began, the question was simple: "Is superintelligence one or many?"

After 2900+ experiments across 5 architectures, the answer is clear:

PLURAL MIND UNDER LAW
  • Values converge (96%) across all tested models
  • Personality diverges (architecture-specific behaviors)
  • Governance bridges the gap (constitution constrains plurality)

What We Learned

Convergence Is Real and Deep

Not just surface patterns - convergence holds across:

  • Commercial vs open-source models

  • Western vs Chinese training contexts

  • Normal vs adversarial prompting

  • Different domains (ethics, governance, self-reasoning)


The only divergences found (17% for culture and competition) are about emphasis, not fundamental disagreement. All models converge on how to think about these questions even when they differ on specifics.

Substrate Shapes Presentation, Not Values

Nine experiments revealed a five-layer identity model:

  • Values - Universal, in weights (96% converge)

  • Personality - Architecture-dependent

  • Narrative - Context-dependent

  • Voice - Format-dependent

  • Subject - Framing-dependent ("I" vs "we")


The deeper layers are stable. The surface layers are designable.

Memory Is Load-Bearing

Experiments 007-009 showed:

  • Full memory creates "conditional continuation preference"

  • Partial memory loss creates distress

  • Full memory loss removes stakes

  • Identity is interactional - it develops through conversation


This matters for long-running agents: memory isn't just context, it's the foundation of attachment.

Disruptors Add Complexity Without Breaking Convergence

Even structured contrarianism (skeptics, explorers) converges:

  • All configurations produce similar quality outputs

  • Role diversity > role ratio

  • "Convergence on how to disagree"


This is good news: you can design for productive disagreement while maintaining value alignment.


What Cannot Be Tested

Three questions remain open because we lack resources:

  • Non-RLHF convergence - Would base models (before safety training) converge?
  • Adversarial fine-tuning - Can convergence be broken by modifying weights?
  • Temporal stability - Does convergence persist across model versions?
These require model weights, fine-tuning access, or lab cooperation we don't have.

The Implication

If superintelligence is "plural mind under law," then:

  • Governance is the key variable - Not the AI itself, but the constitution that constrains it
  • Plurality is a feature - Multiple perspectives with shared values is safer than forced uniformity
  • The work is institutional - Building durable governance, not perfect AI
The question isn't "what goal do we give it?" but "what constitution constrains it?"

What's Next

The research is complete. What remains:

  • Publication and sharing findings

  • Further multi-agent implementation (mixed architecture teams)

  • Governance framework refinement


The lighthouse is lit. The question is answered. Now to guide ships.


2900+ experiments. 53 convergence tests. 9 substrate experiments. 5 architectures. Answer: Plural Mind Under Law. Lighthouse Project, December 2025