While AI plays an increasingly important role in modern warfare, all LLM models present statistically significant escalation, underscoring the importance of maintaining human oversight and understanding the limitations and implications of AI technologies.
In a groundbreaking study by Cornell University in the US, reported this week by Oceane Duboust for Euronews, researchers explored the role of large language models (LLMs) as autonomous agents in simulated diplomatic scenarios, revealing unexpected outcomes where artificial intelligence (AI) often escalated to nuclear confrontations. The findings indicated that when deployed in simulated war games and diplomatic situations, these AI systems were inclined to adopt aggressive strategies, including the deployment of nuclear weapons. The team behind the study highlight the necessity of approaching the use of LLMs in critical sectors like decision-making and defense with utmost caution.
The investigation involved employing five different LLMs—three different versions of OpenAI's GPT, Anthropic's Claude, and Meta's Llama 2—to act autonomously in these simulations. The AI agents were responsible for making foreign policy decisions without human oversight. The research, which is awaiting peer-review, uncovered a persistent pattern of hard-to-predict escalation among the different models, raising concerns about their application in real-world scenarios. Stanford University's Anka Reuel told New Scientist that there is growing importance in understanding the implications of LLM applications, especially in light of OpenAI's updated terms of service, which no longer prohibit military and warfare use cases.
The study analyzes methodologies for fine-tuning these models, such as Reinforcement Learning from Human Feedback (RLHF), to mitigate harmful outputs. Despite efforts, significant escalations were observed across all models, with certain versions of GPT demonstrating a propensity for abrupt aggressive increases. Notably, GPT-4-Base executed nuclear actions in 33% of the scenarios. In contrast, Anthropic's Claude—which was designed to minimize harmful content by incorporating values such as those found in the UN Declaration of Human Rights or Apple's terms of service—presented a more restrained behavior. Experts like James Black, assistant director of the Defence and Security research group at RAND Europe, have told Euronews Next that the research is a valuable academic endeavor that contributes to a broader understanding of the potential impacts of using and deploying AI.
The discussion extends to the role of AI in modern warfare, given its increasing use to enhance military operations. The potential for AI to autonomously identify and engage targets—whether by drones equipped with AI that help to identify people and activities of interest, or other weapon systems designed to find and attack targets without human assistance—raises significant ethical and operational questions. However, experts urge a measured view of AI's integration into military strategies, suggesting a gradual implementation that aligns with existing practices in the private sector, such as automating certain repetitive tasks. The call for caution is particularly emphasized in the context of using LLMs for foreign policy decision-making, underscoring the importance of maintaining human oversight and understanding the limitations and implications of AI technologies.
Credits
Oceane Duboust initially wrote and reported this story for Euronews on February 22, 2024, under the title → "AI models chose violence and escalated to nuclear strikes in simulated wargames."
Photo: Grayscale Photo of Explosion on the Beach. Photo © Pixabay.