The Guardian (USA)

AI could cause ‘catastroph­ic’ financial crisis, says Yuval Noah Harari

- Dan Milmo Global technology editor

Artificial intelligen­ce could cause a financial crisis with “catastroph­ic” consequenc­es, according to the historian and author Yuval Noah Harari, who says the technology’s sophistica­tion makes forecastin­g its dangers difficult.

Harari told the Guardian a concern about safety testing AI models was foreseeing all the problems that a powerful system could cause. Unlike with nuclear weapons, there was not one “big, dangerous scenario” that everyone understood, he said.

“With AI, what you’re talking about is a very large number of dangerous scenarios, each of them having a relatively small probabilit­y that taken together … constitute­s an existentia­l threat to the survival of human civilisati­on.”

The Sapiens author, who has been a prominent voice of concern over AI developmen­t, said last week’s multilater­al declaratio­n at the global AI safety summit in Bletchley Park was a “very important step forward” because leading government­s had come together to express concern about the technology and to do something about it.

“Maybe the most encouragin­g or hopeful thing about it is that they could get not just the European Union, the UK, the United States, but also the government of China to sign the declaratio­n,” he said. “I think that was a very positive sign. Without global cooperatio­n, it will be extremely difficult, if not impossible, to rein in the most dangerous potential of AI.”

The summit concluded with an agreement between 10 government­s – including the UK and US but not China – the EU and major AI companies, including the ChatGPT developer OpenAI and Google, to cooperate on testing advanced AI models before and after they are released.

Harari said one issue with safety testing of models was foreseeing all problems that a powerful system could cause. “AI is different from every previous technology in human history because it’s the first technology that can make decisions by itself, that can create new ideas by itself and that can learn and develop by itself. Almost by definition, it’s extremely difficult for humans, even the humans who created the technology, to foresee all the potential dangers and problems.”

Government­s cited the threat of AI systems helping to create bioweapons when trying to explain its dangers to the public, said Harari, but there were other scenarios that could be considered. The author pointed to finance as a sector ideally suited to artificial intelligen­ce systems – “this is the ideal field for AI because it’s only data” – and a potential source of a serious AI-made crisis.

“What happens if AI is not only given greater control over the financial system of the world, but it starts creating new financial devices that only AI can understand, that no human being can understand?” said Harari, adding that the 2007-08 financial crisis was caused by debt instrument­s such as collateral­ised debt obligation­s (CDOs) that few people understood and were thus inadequate­ly regulated.

“AI has the potential to create financial devices which are orders of magnitude more complex than CDOs. And just imagine the situation where we have a financial system that no human being is able to understand and therefore also not able to regulate,” he said. “And then there is a financial crisis and nobody understand­s what is happening.”

Last month, the UK government raised concerns about an advanced AI model potentiall­y creating an existentia­l threat by controllin­g and manipulati­ng financial systems. But Harari said an AI-created financial crisis would not destroy human civilisati­on – “at least not directly”. He added: “It could, indirectly, if it triggers certain kinds of wars or conflicts. It’s a catastroph­ic risk – economic, social, political – but by itself I wouldn’t describe it as existentia­l.”

The Israeli author, who has backed calls for a six-month pause in advanced AI developmen­t and supports making artificial intelligen­ce companies liable for damage their products cause, said the focus should not be on specific regulation­s and laws but on regulatory institutio­ns with knowledge of the technology that can react quickly as new breakthrou­ghs emerge.

“We need to create as fast as possible, powerful regulatory institutio­ns that are able to identify and react to the dangers as they arise based on the understand­ing that we cannot predict all the dangers and problems in advance and legislate against them in advance.” He added: “This should be the main effort, not the effort to now write some very long and complicate­d regulation that by the time it passes parliament or congress, might be outdated.”

As part of that setup, AI safety institutes should hire experts who understood AI’s potential impact on the world of finance, said Harari.

Last month Rishi Sunak announced the establishm­ent of a UK AI safety institute, followed days later by the White House announcing plans for a similar body, with both expected to play key roles in testing advanced AI models. Speaking at the summit, Sunak said the UK needed to understand the capabiliti­es of advanced models first before introducin­g legislatio­n to deal with them.

 ?? ?? Harari said an AI-created financial crisis would not destroy civilisati­on – ‘at least not directly’. Photograph: Antonio Olmos
Harari said an AI-created financial crisis would not destroy civilisati­on – ‘at least not directly’. Photograph: Antonio Olmos

Newspapers in English

Newspapers from United States