AI & ML Scaling Insight

This paper establishes the formal information-theoretic limits and conditions under which self-improving AI systems can be safely verified.

March 31, 2026

Original Paper

Information-Theoretic Limits of Safety Verification for Self-Improving Systems

Arsenios Scrivens

arXiv · 2603.28650

The Takeaway

It provides a rigorous mathematical proof that safety gates based on fixed classifiers will inevitably fail as systems improve, but also identifies a 'safety-exit' via Lipschitz-bounded conditional logic that allows for safe unbounded self-modification.

From the abstract

Can a safety gate permit unbounded beneficial self-modification while maintaining bounded cumulative risk? We formalize this question through dual conditions -- requiring sum delta_n 1, any classifier-based gate under overlapping safe/unsafe distributions satisfies TPR_n 0, escaping the impossibility. Formal Lipschitz bounds for pre-LayerNorm transformers under LoRA enable LLM-scale verification. The separation is strict. We validate on GPT-2 (d_LoRA = 147,456): conditional delta = 0 with TPR =