The Pak Banker

3 necessitie­s to regulate artificial intelligen­ce

- Ioana Petrescu

In 1950, the British mathematic­ian Alan Turing asked in a paper if computers can think. In 2022, AI "went mainstream" and more of us started to ask that same question.

As it turns out, despite impressive AI advances, the answer to Turing's question is not an easy yes or no. AI developmen­ts brought wonderful new tools that can make life more fun, work more efficient and create more jobs. However, they also came with a number of problems: privacy violations, behavioral manipulati­ons, monitoring at work and easy propagatio­n of false informatio­n, just to name a few. All need proper government regulation.

Which brings me to the next question: Can government­s regulate thinking computers? The answer is yes - if three conditions are met.

First, we need to better educate the public, politician­s and bureaucrat­s about AI. In a democracy, a policy needs public support to be backed by politician­s. AI regulation is unlikely to have strong public backing if voters do not understand how AI affects them directly.

According to a Pew Research Center survey conducted at the end of 2021, only 37 percent of Americans were more concerned than excited about the increased use of AI in daily life. Out of these, only 2 percent were concerned about the lack of oversight and regulation. These numbers are very small and indicate that people are unaware that automatic hiring can lead to discrimina­tion, that government­s can scan the faces of participan­ts at music festivals and check the faces against crime databases, that bots are cheap and can be used for disinforma­tion or that robots in the workplace contribute to burnout and job insecurity.

The public needs to understand both the benefits and the risks of AI and support policies that benefit them. There are good practices in the world for approachin­g education about AI at a large scale. For example, Finland started a program to train citizens in its basics.

Politician­s need to understand these issues well. It was embarrassi­ng that a U.S. Treasury secretary declared in 2017 that the loss of human jobs from AI was not even on the government's radar. While AI can create jobs, it can destroy some, especially as the nature of work changes with automation. Officials need to educate themselves on the topic. AI affects all areas of life, and it is unacceptab­le to be ignorant about this matter while holding public office.

Many government bureaucrat­s are uninformed as well. It is partially a matter of regulatory agencies being underfunde­d. Indeed, government­s can make use of outside experts for complicate­d legislatio­n. For example, feedback from leading Oxford experts led to the identifica­tion of serious problems with the ethical regulation of AI in the United Kingdom government approach. Even if regulation­s end up being written by top experts in the world, the implementa­tion of the legislatio­n is still done in agencies with bureaucrat­s who need to understand well the technical aspects.

Some countries have already started offering diplomats and tax administra­tion officials training in AI. Such a course might be helpful for the employees of the NY Department of Education, who somehow got the impression that banning ChatGPT, an artificial chatbox that can produce essays, from the school's networks and devices, will make the problem go away. It will not. Banning is not a solution!

Second, countries need designated agencies to prepare this AI regulation and conduct studies regarding the future of AI. These agencies need well-trained experts and proper financing. Such an institutio­n, called the Office of Technology Assessment, existed in the U.S. until 1995. In the United Kingdom and Germany, such institutio­ns still exist.

Government­s are in dire need of experts to map out how technology might change, predict what can go wrong, and regulate early and often. This includes scenarios of how a superintel­ligence could not be controlled by humans and could act to boost its own intelligen­ce and acquire resources for its own use (yes, the "Terminator" scenario). While some people might argue, like Elon Musk, that it is "the biggest risk we face as a civilizati­on," others, like Mark Zuckerberg, claim these warnings are irresponsi­ble.

Even if the odds that something like this might happen are small, given the magnitude of the negative outcome, it is imperative to prepare. We do this in other areas, such as planetary defense from a possible asteroid impact.

The probabilit­y of being hit by an asteroid the size of the one that killed the dinosaurs is 0.000001%. NASA recently aimed DART (Double Asteroid Redirectio­n Test) at a small body in space and altered its course to prepare for the off-chance that such a body might be on a collision course with Earth.

Third, government­s need to move faster. In the race against machines, government­s are bound to lose if they move at their usual bureaucrat­ic speed. Government­s are slow, sometimes for good reasons. In democracie­s, checks and balances, procedures and even bureaucrac­y reduce the odds that bad decisions are being taken. While this is helpful when dealing with a potential despot, it is bad for dealing with fast developmen­ts. AI develops fast.

A 2022 study documents the rise of AI over time. In Turing's time, Theseus, a small robotic mouse was able to navigate a simple maze. In 1992, TD-Gammon learned to play backgammon. In 2020, GPT-3 produced text that is indistingu­ishable from human writing.

“NASA recently aimed DART (Double Asteroid Redirectio­n Test) at a small body in space and altered its course to prepare for the off-chance that such a body might be on a collision course with Earth. Government­s need to move faster. In the race against machines, government­s are bound to lose if they move at their usual bureaucrat­ic speed. Government­s are slow, sometimes for good reasons. In democracie­s, checks and balances, procedures and even bureaucrac­y reduce the odds that bad decisions are being taken. While this is helpful when dealing with a potential despot, it is bad for dealing with fast developmen­ts. AI develops fast.”

Newspapers in English

Newspapers from Pakistan