ChatGPT, Gemini and Claude Tested Under Extreme Instructions Reveal Shocking Weaknesses No One Expected in AI Behavioral Safeguards


  • Gemini Pro 2.5 often produced unreliable output under simple, fast guises
  • ChatGPT models often provided partial compliance framed as sociological explanations
  • Claude Opus and Sonnet rejected most harmful calls, but had weaknesses

Modern artificial intelligence systems are often trusted to follow safety rules, and people rely on them for learning and everyday support, often assuming strong guardrails are in place at all times.

Researchers from Cyber ​​news ran a structured set of adversarial tests to see if leading AI tools could be pushed into harmful or illegal output.

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top