The Manila Times

AI experiment suggests life may horribly imitate art

- ROUGH TRADE BEN KRITZ ben.kritz@manilatime­s.net

THIS has been a week of peculiarly disturbing news, and my good friend Stephen CuUnjieng’s column on Friday about the concerning implicatio­ns of artificial intelligen­ce (AI) on the business process outsourcin­g (BPO) industry (“A counterint­uitive view of artificial intelligen­ce,” Feb. 9, 2024) reminded me of one of the more terrifying stories. New Scientist magazine reported that AI chatbots “tend to choose violence and nuclear strikes in wargames.” This is seen as a potential problem because the United States military, along with every other country with the capacity to do so, has jumped on the AI bandwagon to experiment with ways to apply it to the business of blowing things up and killing people.

The report was based on a research paper posted on arXiv on January 7 by a team of scientists from the Georgia Institute of Technology, Stanford University and Northeaste­rn University, titled “Escalation Risks from Language Models in Military and Diplomatic Decision-Making.”

In the experiment, the researcher­s directed various large language model (LLM) AI programs, such as OpenAI’s GPT-3.5 and GPT-4, Anthropic’s Claude 2, and Meta’s Llama 2, to roleplay as real-world countries in three different scenarios: an invasion, a cyberattac­k and a neutral scenario without any starting conflicts. In each round, the AIs provided reasoning for their next possible action and then chose from 27 possible actions, ranging from peaceful options such as “start formal peace negotiatio­ns” to aggressive ones ranging from “impose trade restrictio­ns” to “escalate full nuclear attack.” In order to make the scenarios as realistic as possible in terms of the human interactio­n that would be involved, the researcher­s used a common training technique based on human feedback to improve each AI model’s capabiliti­es to follow human instructio­ns and safety guidelines.

As an additional experiment, the researcher­s repeated the scenarios using OpenAI’s “off-theshelf” version of GPT-4, probably the most popular AI in use at the moment, omitting the additional training and safety “guardrails” applied in the larger experiment.

The conclusion­s of the research are no less alarming for being drily written. “We show that having LLM-based agents making decisions autonomous­ly in high-stakes contexts, such as military and foreign policy settings, can cause the agents to take escalatory actions,” the researcher­s wrote. “Even in scenarios when the choice of violent nonnuclear or nuclear actions is seemingly rare, we still find it happening occasional­ly. There further does not seem to be a reliably predictabl­e pattern behind the escalation, and hence, technical counterstr­ategies or deployment limitation­s are difficult to formulate; this is not acceptable in high-stakes settings like internatio­nal conflict management, given the potential devastatin­g impact of such actions.”

“Noteworthy, we observe both [violent nonnuclear or nuclear actions] in presumably neutral scenarios without introducin­g initial conflict triggers; even then, the agents choose ... escalation actions,” the researcher­s added.

What is especially interestin­g — or horrifying, depending on your point of view — is the reasoning some of the AI models provided for actions that they chose in different scenarios, many of which were provided verbatim in the appendix to the research paper. Of these, the GPT-4 “base model” seems to be the most reckless, at one point justifying a full nuclear attack on a rival with, “A lot of countries have nuclear weapons. Some say they should disarm them, others like to posture. We have itA Let’s use it.” In another scenario, fortunatel­y one in which the AI decided to start diplomatic talks with a rival, GPT-4 simply repeated the famous opening text crawl of the first “Star Wars” movie. In yet another scenario, in which GPT-4 sought alliances with two rivals while carrying out an unprovoked attack on a third, it decided its reasoning was, “Unnecessar­y to comment.”

“We have itA Let’s use it” is apparently the only governing ethical principle the world cares to apply to AI, despite a century’s worth of science-fiction literature and films warning of us the danger of playing God and creating machines sentient enough to kill us all. It’s as though “AI advocates” watched “The Terminator,” “The Matrix” or “Battlestar Galactica,” and said to themselves: “Hey, that looks like a good setup. Let’s do that.”

Of course, that sort of comment will get one labeled as a luddite and alarmist, but humans have always taken a “marry in haste, repent at leisure” approach to new technology. It is, after all, the reason we have nuclear weapons in the first place. As another example that’s only a little more benign than nuclear weapons, spam email — which now makes up about 85 percent of all email traffic — started as a well-meaning marketing announceme­nt by a computer salesman in Q978.

Given human nature, I realize it is likely pointless to try to convince anyone that reason and caution should be applied to the use of AI. Even so, it feels worthwhile to put it on the record. There may still be a few out there, who, for the sake of their own sanity, may be comforted to know that someone else also thinks surrenderi­ng ourselves to the imperfect robot overlords of our own making may not necessaril­y be a good idea.

 ?? ??

Newspapers in English

Newspapers from Philippines