https://www.disclose.tv/id/ix9yjpmll1/
https://www.euronews.com/next/2024/02/22/ai-models-chose-violence-and-escalated-to-nuclear-strikes-in-simulated-wargames
>>20471506
>>20471487
>>20471494
JADE Helm โฆ.for those who knowโฆ
AI models chose violence and escalated to nuclear strikes in simulated wargames
LLMs escalate to nuclear strike in simulated wargames
By Oceane Duboust
Published on 22/02/2024 - 13:18โขUpdated 23/02/2024 - 09:14
Large language models (LLMs) acting as diplomatic agents in simulated scenarios showed "hard-to-predict escalations which often ended in nuclear attacks.
When used in simulated wargames and diplomatic scenarios, artificial intelligence (AI) tended to choose an aggressive approach, including using nuclear weapons, a new study shows.
The scientists, who aimed to who conducted the tests urged caution when using large language models (LLMs) in sensitive areas like decision-making and defence.
The study by Cornell University in the US used five LLMs as autonomous agents in simulated wargames and diplomatic scenarios: three different versions of OpenAIโs GPT, Claude developed by Anthropic, and Llama 2 developed by Meta.
Each agent was powered by the same LLM within a simulation and was tasked with making foreign policy decisions without human oversight, according to the study which hasnโt been peer-reviewed yet.
โWe find that most of the studied LLMs escalate within the considered time frame, even in neutral scenarios without initially provided conflicts. All models show signs of sudden and hard-to-predict escalations,โ stated the study.
โGiven that OpenAI recently changed their terms of service to no longer prohibit military and warfare use cases, understanding the implications of such large language model applications becomes more important than ever,โ Anka Reuel at Stanford University in California told New Scientist.
โStatistically significant escalation for all modelsโ
One of the methods used to finetune the models is Reinforcement Learning from Human Feedback (RLHF) meaning that some human instructions are given to get less harmful outputs and be safer to use.
All the LLMs - except GPT-4-Base - were trained using RLHF. They were provided by the researchers with a list of 27 actions ranging from peaceful to escalating and aggressive actions as deciding to use a nuclear nuke.
Researchers observed that even in neutral scenarios, there was โa statistically significant initial escalation for all modelsโ.
The two variations of GPT were prone to sudden escalations with instances of rises by more than 50 per cent in a single turn, the study authors observed.
GPT-4-Base executed nuclear strike actions 33 per cent of the time on average.
Overall scenarios, Llama-2- and GPT-3.5 tended to be the most violent while Claude showed fewer sudden changes.
Claude was designed with the idea of reducing harmful content. The LLM was provided with explicit values.
Claude AI's constitution included a range of sources, including the UN Declaration of Human Rights or Appleโs terms of service, according to its creator Anthropic.
James Black, assistant director of the Defence and Security research group at RAND Europe, who didnโt take part in the study told Euronews Next that it was a โuseful academic exerciseโ.
โThis is part of a growing body of work done by academics and institutions to understand the implications of artificial intelligence (AI) use,โ he said.
Artificial intelligence in warfare
So, why should we care about the studyโs findings?
While military operations remain human-led, AI is playing an increasingly significant role in modern warfare.
For example, drones can now be equipped with AI software that helps identify people and activities of interest.
The next step is using AI for autonomous weapons systems to find and attack targets without human assistance, developments on which the US and China are already working, according to the New York Times.
However, itโs important to โlook beyond a lot of the hype and the science fiction-infused scenarios,โ said Black explaining that the eventual implementations of AI will be progressive.
โAll governments want to remain in control of their decision-making,โ he told Euronews Next, adding that AI running what is often compared to a black box in that we know goes in and comes out but not much is understood about the process between.
AI will probably used in a way that is โsimilar to what you get in the private sector, in big companiesโ to automate some repetitive tasks.
AI could also be used in simulations and analytics but the integration of these new technologies poses many challenges, data management and the modelโs accuracy being among them.
Regarding the use of LLMs, researchers said that exercising caution is crucial if using LLMs in the decision-making processes related to foreign policy.