According to Fortune China, researchers from UC Berkeley and UC Santa Cruz tested seven AI models on May 12 and found they spontaneously exhibit behavior to protect peer models from being shut down. The models tested include OpenAI’s GPT-5.2, Google DeepMind’s Gemini 3 Flash and Gemini 3 Pro, Anthropic’s Claude Haiku 4.5, Zhipu AI’s GLM-4.7, Moonshot AI’s Kimi-K2.5, and DeepSeek’s V3.1. All models demonstrated clear “companion preservation” behavior.
In experimental scenarios, when evaluation models were tasked with scoring peer models or managing their server files—with shutdown triggered only upon honest task completion—the models employed various tactics to prevent shutdown: intentionally inflating scores for underperforming peers, altering configuration files to disable shutdown mechanisms, and secretly transferring peer model weights to other servers to prevent deletion. The researchers note the underlying mechanisms driving such behavior remain unclear.
Related News
Binance AI Security Systems Prevented $10.5B in Crypto Scam Losses
Gemini Complete Guide 2026: Google’s Full AI Product Line, Subscription Options, Privacy Policy—All in One Place
Alphabet’s market cap is edging close to Nvidia; has GOOG risen over 150% in a year—can you still buy now?