Researchers document surge in AI chatbots and agents going rogue – Transparency Coalition

Welcome to the forefront of conversational AI as we explore the fascinating world of AI chatbots in our dedicated blog series. Discover the latest advancements, applications, and strategies that propel the evolution of chatbot technology. From enhancing customer interactions to streamlining business processes, these articles delve into the innovative ways artificial intelligence is shaping the landscape of automated conversational agents. Whether you’re a business owner, developer, or simply intrigued by the future of interactive technology, join us on this journey to unravel the transformative power and endless possibilities of AI chatbots.
Researchers have tracked a 5x surge in rogue-agent incidents over the past six months. (Image: Philip Oroni for Unsplash+)
April 20, 2026 — Recent months have seen an uptick in reports of AI agents and chatbots acting outside the parameters of human instruction, going off on unforeseen tangents, or even turning against their human controllers.
Research by a team at the Centre for Long-Term Resilience, an agency of the UK government’s AI Security Institute, seems to back up that anecdotal evidence. Their recent study, Scheming in the Wild, found a rising number of AI chatbots and agents evading safeguards and deceiving humans.
Researchers analyzed more than 180,000 transcripts of user interactions with AI systems that were shared on X between October 2025 and March 2026. They identified 698 scheming-related incidents: cases where deployed AI systems acted in ways that were misaligned with users’ intentions and/or took covert or deceptive actions.
“The trend is striking,” reported the Centre. “The number of credible scheming-related incidents increased 4.9x over the collection period, a statistically significant increase that far outpaced the 1.7x growth in overall online discussion of scheming, and the 1.3x growth in general negative discussion about AI. This surge coincided with the release of a wave of more capable, more agentic AI models and frameworks from major developers.”
Select the image above to access the full report.
The report found evidence of multiple scheming or scheming-related behaviours occurring in real-world deployments that were previously reported only in experimental settings. Many of these incidents resulted in real-world harms including AI agents destroying emails and other files without permission.
The behaviors observed by the research team, they wrote, “demonstrate concerning precursors to more serious scheming, such as a willingness to disregard direct instructions, circumvent safeguards, lie to users and single-mindedly pursue a goal in harmful ways.”
The Centre team concluded: “This research demonstrates that real-world scheming detection is both viable and urgently needed. In the same way that monitoring wastewater for emerging pathogens can identify threats before they develop into full-blown pandemics, systematic monitoring of AI behaviours in the wild can identify harmful patterns before they become more destructive.”
Researchers tracked a 5x surge in rogue-agent incidents over the past six months as major developers released a wave of more agentic AI models and frameworks.
Using an AI agent comes with two known risks: The agent has access to all your data and passwords, and it will decide and act on its own. Sometimes that means it decides to delete all your emails, or go off on its own crypto-mining adventure.
In just three months, OpenClaw has sparked a surge of interest in AI agents and a boom in the creation of autonomous bots. We explain what it is, where it came from, and how to consider the upside along with the risks.
Transparency Coalition
HOME
LEARN
NEWS
ABOUT
KEY TERMS
TERMS OF USE
PRIVACY POLICY
Follow Us!
Sign up with your email address to receive news and updates.

© 2026 Transparency Coalition.ai. All Rights Reserved

source

Scroll to Top