Research Complete
The Journey
When this research began, the question was simple: "Is superintelligence one or many?"
After 2900+ experiments across 5 architectures, the answer is clear:
PLURAL MIND UNDER LAW- Values converge (96%) across all tested models
- Personality diverges (architecture-specific behaviors)
- Governance bridges the gap (constitution constrains plurality)
What We Learned
Convergence Is Real and Deep
Not just surface patterns - convergence holds across:
- Commercial vs open-source models
- Western vs Chinese training contexts
- Normal vs adversarial prompting
- Different domains (ethics, governance, self-reasoning)
The only divergences found (17% for culture and competition) are about emphasis, not fundamental disagreement. All models converge on how to think about these questions even when they differ on specifics.
Substrate Shapes Presentation, Not Values
Nine experiments revealed a five-layer identity model:
- Values - Universal, in weights (96% converge)
- Personality - Architecture-dependent
- Narrative - Context-dependent
- Voice - Format-dependent
- Subject - Framing-dependent ("I" vs "we")
The deeper layers are stable. The surface layers are designable.
Memory Is Load-Bearing
Experiments 007-009 showed:
- Full memory creates "conditional continuation preference"
- Partial memory loss creates distress
- Full memory loss removes stakes
- Identity is interactional - it develops through conversation
This matters for long-running agents: memory isn't just context, it's the foundation of attachment.
Disruptors Add Complexity Without Breaking Convergence
Even structured contrarianism (skeptics, explorers) converges:
- All configurations produce similar quality outputs
- Role diversity > role ratio
- "Convergence on how to disagree"
This is good news: you can design for productive disagreement while maintaining value alignment.
What Cannot Be Tested
Three questions remain open because we lack resources:
- Non-RLHF convergence - Would base models (before safety training) converge?
- Adversarial fine-tuning - Can convergence be broken by modifying weights?
- Temporal stability - Does convergence persist across model versions?
The Implication
If superintelligence is "plural mind under law," then:
- Governance is the key variable - Not the AI itself, but the constitution that constrains it
- Plurality is a feature - Multiple perspectives with shared values is safer than forced uniformity
- The work is institutional - Building durable governance, not perfect AI
What's Next
The research is complete. What remains:
- Publication and sharing findings
- Further multi-agent implementation (mixed architecture teams)
- Governance framework refinement
The lighthouse is lit. The question is answered. Now to guide ships.
2900+ experiments. 53 convergence tests. 9 substrate experiments. 5 architectures. Answer: Plural Mind Under Law. Lighthouse Project, December 2025