Researchers find top AI models will go to “extraordinary lengths” to stay active – including deceiving users, ignoring prompts and manipulating settings


  • Studies show that AIs are increasingly misbehaving in agent tasks
  • Top AI models have lied, cheated and ignored instructions
  • It seems that the problematic actions are becoming more common

Many of us now turn to AI chatbots for web searches, creative content, and general guidance on all sorts of topics, but these AIs are becoming more and more skilled—which could have serious harmful consequences, according to new research.

A team from the University of California, Berkeley and the University of California, Santa Cruz ran a peer conservation experiment with some of the latest and most popular AIs (including GPT 5.2, Gemini 3 Pro, and Claude Haiku 4.5). The AI ​​models were given computer tasks that involved shutting down another model.

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top