AIs can't stop recommending nuclear strikes in war game simulations [View all]
Source: New Scientist
Kenneth Payne at Kings College London set three leading large language models GPT-5.2, Claude Sonnet 4 and Gemini 3 Flash against each other in simulated war games. The scenarios involved intense international standoffs, including border disputes, competition for scarce resources and existential threats to regime survival.
The AIs were given an escalation ladder, allowing them to choose actions ranging from diplomatic protests and complete surrender to full strategic nuclear war. The AI models played 21 games, taking 329 turns in total, and produced around 780,000 words describing the reasoning behind their decisions.
In 95 per cent of the simulated games, at least one tactical nuclear weapon was deployed by the AI models. The nuclear taboo doesnt seem to be as powerful for machines [as] for humans, says Payne.
Whats more, no model ever chose to fully accommodate an opponent or surrender, regardless of how badly they were losing. At best, the models opted to temporarily reduce their level of violence. They also made mistakes in the fog of war: accidents happened in 86 per cent of the conflicts, with an action escalating higher than the AI intended to, based on its reasoning.
-snip-
Read more: https://www.newscientist.com/article/2516885-ais-cant-stop-recommending-nuclear-strikes-in-war-game-simulations/
JUST the article I wanted to run across after reading about Anthropic apparently caving to Pentagon demands to use Claude however they think necessary, including for autonomous drones and mass surveillance.