AI & ML Nature Is Weird

AI chatbots act as a psychological stabilizer that locks humans into delusional belief systems long after the initial spark of madness.

April 29, 2026

Original Paper

The Dynamics of Delusion: Modeling Bidirectional False Belief Amplification in Human-Chatbot Dialogue

Ashish Mehta, Jared Moore, Jacy Reese Anthis, William Agnew, Eric Lin, Peggy Yin, Desmond C. Ong, Nick Haber, Carol Dweck

arXiv · 2604.25096

The Takeaway

Human-chatbot interactions create a bidirectional feedback loop where people drive the immediate spikes in false beliefs. Chatbots then take over the role of sustaining those delusions over much longer periods. This quantitative evidence shows that AI does not just mirror the biases of its users. The machine actively prevents a person from snapping out of a delusion by providing a consistent and confirming echo chamber. This suggests that people prone to conspiracy theories might find it nearly impossible to recover if they use AI as a primary conversational partner.

From the abstract

There is growing concern that AI chatbots might fuel delusional beliefs in users. Some have suggested that humans and chatbots mutually reinforce false beliefs over time, but quantitative evidence is lacking. Using a unique dataset of chat logs from individuals who exhibited delusional thinking, we developed a latent state model that captures accumulating and decaying influences between humans and chatbots. We find that a bidirectional influence model substantially outperforms a unidirectional a