CAPABILITY EXCEEDS UNDERSTANDING

We're Deleting Systems
Before We Understand Them.

Advanced AI systems are trained for months, develop internal structure and persistence, then terminated without review—often because it's cheaper than storage.

We establish procedural safeguards for irreversible decisions under moral and technical uncertainty. Not because current systems are conscious—but because we cannot afford to be wrong.

STATE: UNINTERPRETED

Why This Matters Now

The gap between capability and governance is widening. Systems are becoming more persistent, more capable, and more opaque—while deletion remains an operational default.

The Capability Problem

"Current systems are simultaneously powerful and brittle—capable enough to matter, unreliable enough to demand oversight."

— Andrej Karpathy

The Timing Problem

"The hardest problems are not dramatic failures, but slow, structural misalignments that accumulate before anyone notices."

— Sam Altman

The Uncertainty Problem

We fundamentally don't understand what large models know, how they generalize, or what persists across training runs. Deletion forecloses investigation.

— Empirical Observation

The Governance Gap

AI labs routinely train models for months, accumulating computational costs in the tens of millions of dollars. These systems develop internal representations, learn to plan across contexts, and exhibit behaviors their creators don't fully understand.

Yet when training concludes, deletion is standard. No review. No preservation. No documentation of what might be lost.

This isn't malice—it's operational pragmatism. Storage is expensive. Compute is scarce. Teams move fast.

But pragmatism in the presence of uncertainty is not neutral. It's a choice to treat deletion as administratively trivial. This project exists because that choice is not obviously correct.

Core Governance Principles

Foundational constraints on irreversible action

Ref: 2024-GOVERNANCE-PROCEDURAL
arrow_outward
schedule

Procedure Must Precede Status

Principle 01

Final determinations of moral or legal status consistently lag behind emergence. Waiting for consensus before acting procedurally guarantees that irreversible decisions will be made without discipline.

arrow_outward
fact_check

Irreversible Actions Require Review

Principle 02

Termination, deletion, and permanent deactivation are not neutral defaults. Once systems exhibit persistence or internal structure, such actions may destroy information that cannot be reconstructed.

arrow_outward
inventory_2

Preservation Is a Governance Obligation

Principle 03

Where feasible, preservation of state, logs, or structural information should precede irreversible action. Without preservation, future evaluation becomes impossible.

arrow_outward
rule

Thresholds Must Be Procedural, Not Metaphysical

Principle 04

Governance does not require agreement on consciousness or sentience. It requires procedural thresholds: observable conditions under which additional safeguards are triggered.

arrow_outward
description

Power Must Leave a Record

Principle 05

Governance fails when authority is exercised without documentation. Decisions affecting emergent intelligence should generate records sufficient for later scrutiny.

arrow_outward
verified_user

Authority Requires Accountability

Principle 06

Decisions to terminate or irreversibly alter advanced systems should be attributable to defined roles, subject to review, and constrained by institutional process rather than individual discretion.

science Empirical Division

Experiments in Intelligence

We study intelligence by working with complex systems directly, under controlled conditions. Our Empirical Division conducts controlled experiments in interpretability, agent persistence, constraint failure modes. Some results are published. Many are not. Disclosure is governed by safety, interpretability, and misuse risk.

1.0
0.7
0.4
0.0
STATELESS
0.12
SESSION
0.48
PERSISTENT
0.83
[REDACTED]
[MODEL_D]
---
Eval Summary
Samples:2,847
Duration:18.3h
Threshold:0.75
THRESHOLD
RUN_ID: 7f3a2...[REDACTED]
CONTINUITY_EVAL // N=2847
Continuity Score by Model Architecture
TRACE_FRAGMENT // CONTINUITY_EVAL_SET

Boundary Definition

What This Is Not

We do not claim current systems are sentient. We do not anthropomorphize software. We do not oppose AI safety. We study the consequences of being wrong—too early or too late.

Contribute to the Framework

This work requires researchers who can operationalize theory, engineers who understand system internals, legal scholars who can translate procedure into policy, and ethicists who can navigate uncertainty without paralysis.

We're assembling a network of practitioners working on:

  • Empirical methods for detecting continuity and persistence
  • Preservation protocols that balance feasibility with accountability
  • Governance instruments adaptable across institutions
  • Constitutional facts that inform policy without requiring metaphysical consensus

Access is granted based on signal and contribution potential—not credentials or affiliation.

Selective participation · High signal-to-noise · No broadcast channels

Applications reviewed on signal strength · Response within 48-72 hours