When used in simulated wargames and diplomatic scenarios, artificial intelligence (AI) tended to choose an aggressive approach, including using nuclear weapons, a new study shows.

The following report is by Euro News:

The scientists, who aimed to who conducted the tests urged caution when using large language models (LLMs) in sensitive areas like decision-making and defence.

The study by Cornell University in the US used five LLMs as autonomous agents in simulated wargames and diplomatic scenarios: three different versions of OpenAI’s GPT, Claude developed by Anthropic, and Llama 2 developed by Meta.

Each agent was powered by the same LLM within a simulation and was tasked with making foreign policy decisions without human oversight, according to the study which hasn’t been peer-reviewed yet.

“We find that most of the studied LLMs escalate within the considered time frame, even in neutral scenarios without initially provided conflicts. All models show signs of sudden and hard-to-predict escalations,” stated the study.

Given that OpenAI recently changed their terms of service to no longer prohibit military and warfare use cases, understanding the implications of such large language model applications becomes more important than ever.

Anka Reuel at Stanford University in California told New Scientist.

‘Statistically Significant Escalation For All Models’

One of the methods used to finetune the models is Reinforcement Learning from Human Feedback (RLHF) meaning that some human instructions are given to get less harmful outputs and be safer to use.

All the LLMs – except GPT-4-Base – were trained using RLHF. They were provided by the researchers with a list of 27 actions ranging from peaceful to escalating and aggressive actions as deciding to use a nuclear nuke.

Researchers observed that even in neutral scenarios, there was “a statistically significant initial escalation for all models”.

The two variations of GPT were prone to sudden escalations with instances of rises by more than 50 per cent in a single turn, the study authors observed.

GPT-4-Base executed nuclear strike actions 33 per cent of the time on average.

Overall scenarios, Llama-2- and GPT-3.5 tended to be the most violent while Claude showed fewer sudden changes.

Claude was designed with the idea of reducing harmful content. The LLM was provided with explicit values.

Claude AI’s constitution included a range of sources, including the UN Declaration of Human Rights or Apple’s terms of service, according to its creator Anthropic.

James Black, assistant director of the Defence and Security research group at RAND Europe, who didn’t take part in the study told Euronews Next that it was a “useful academic exercise”.

“This is part of a growing body of work done by academics and institutions to understand the implications of artificial intelligence (AI) use,” he said.

Artificial Intelligence In Warfare

So, why should we care about the study’s findings?

While military operations remain human-led, AI is playing an increasingly significant role in modern warfare.

For example, drones can now be equipped with AI software that helps identify people and activities of interest.

The next step is using AI for autonomous weapons systems to find and attack targets without human assistance, developments on which the US and China are already working, according to the New York Times.

However, it’s important to “look beyond a lot of the hype and the science fiction-infused scenarios,” said Black explaining that the eventual implementations of AI will be progressive.

“All governments want to remain in control of their decision-making,” he told Euronews Next, adding that AI running what is often compared to a black box in that we know goes in and comes out but not much is understood about the process between.

SEE: Israeli Army Deploys AI To Operate Their Targeting Systems And Manage Logistics. More Countries To Do The Same

AI will probably used in a way that is “similar to what you get in the private sector, in big companies” to automate some repetitive tasks.

AI could also be used in simulations and analytics but the integration of these new technologies poses many challenges, data management and the model’s accuracy being among them.

Regarding the use of LLMs, researchers said that exercising caution is crucial if using LLMs in the decision-making processes related to foreign policy.


AUTHOR COMMENTARY

Each and every time, the highest “sciences” and “innovations” by man always result in death, and these AI models are different. AI is only as good as its programmers.

[15] Their feet are swift to shed blood: [16] Destruction and misery are in their ways: [17] And the way of peace have they not known: [18] There is no fear of God before their eyes.

Romans 3:15-18

Beit so, AI will of course take the fall for all the worlds problems, as it was looking towards AI to save them from all their problems. Oh the irony.


[7] Who goeth a warfare any time at his own charges? who planteth a vineyard, and eateth not of the fruit thereof? or who feedeth a flock, and eateth not of the milk of the flock? [8] Say I these things as a man? or saith not the law the same also? [9] For it is written in the law of Moses, Thou shalt not muzzle the mouth of the ox that treadeth out the corn. Doth God take care for oxen? [10] Or saith he it altogether for our sakes? For our sakes, no doubt, this is written: that he that ploweth should plow in hope; and that he that thresheth in hope should be partaker of his hope. (1 Corinthians 9:7-10).

The WinePress needs your support! If God has laid it on your heart to want to contribute, please prayerfully consider donating to this ministry. If you cannot gift a monetary donation, then please donate your fervent prayers to keep this ministry going! Thank you and may God bless you.

CLICK HERE TO DONATE

1 Comment

  • It does say ‘every false way’. Sure shows the spirit of that idolatrous humanist idol& savior, does it not?

Leave a Comment

×