AI & ML Paradigm Shift

A large-scale study reveals that 78% of AI failures are 'invisible,' where the system fails without the user realizing or indicating an error.

March 17, 2026

Original Paper

Invisible failures in human-AI interactions

Christopher Potts, Moritz Sudhof

arXiv · 2603.15423

The Takeaway

This shifts the focus of AI safety and monitoring from visible user complaints to latent system errors. It introduces an eight-archetype taxonomy for these failures, suggesting that standard feedback loops are insufficient for building reliable production systems.

From the abstract

AI systems fail silently far more often than they fail visibly. In a large-scale quantitative analysis of human-AI interactions from the WildChat dataset, we find that 78% of AI failures are invisible: something went wrong but the user gave no overt indication that there was a problem. These invisible failures cluster into eight archetypes that help us characterize where and how AI systems are failing to meet users' needs. In addition, the archetypes show systematic co-occurrence patterns indica