Jamaica Gleaner

AI in the common interest

-

THE TECH world has generated a fresh abundance of frontpage news in 2022.

In October, Elon Musk bought Twitter – one of the main public communicat­ion platforms used by journalist­s, academics, businesses, and policymake­rs – and proceeded to fire most of its content-moderation staff, indicating that the company would rely instead on artificial intelligen­ce, or AI.

Then, in November, a group of Meta employees revealed that they had devised an AI program capable of beating most humans in the strategy game Diplomacy.

In Shenzhen, China, officials are using “digital twins” of thousands of 5G-connected mobile devices to monitor and manage flows of people, traffic, and energy consumptio­n in real time. And with the latest iteration of Chatgpt’s language-prediction model, many are declaring the end of the college essay.

In short, it was a year in which already-serious concerns about how technologi­es are being designed and used deepened into even more urgent misgivings. Who is in charge here? Who should be in charge?

Public policies and institutio­ns should be designed to ensure that innovation­s are improving the world, yet many technologi­es are currently being deployed in a vacuum. We need inclusive mission-oriented governance structures that are centred around a true common good. Capable government­s can shape this technologi­cal revolution to serve the public interest.

AI, which the Oxford English Dictionary defines broadly as “the theory and developmen­t of computer systems able to perform tasks normally requiring human intelligen­ce, such as visual perception, speech recognitio­n, decision-making, and translatio­n between languages”, can make our lives better in many ways. It can enhance food production and management, by making farming more efficient and improving food safety. It can help us bolster resilience against natural disasters, design energy-efficient buildings, improve power storage and optimise renewable energy deployment. And it can enhance the accuracy of medical diagnostic­s when combined with doctors’ own assessment­s.

These applicatio­ns would make our lives better in many ways.

But with no effective rules in place, AI is likely to create new inequaliti­es and amplify pre-existing ones. One need not look far to find examples of Ai-powered systems reproducin­g unfair social biases. In one recent experiment, robots powered by a machine-learning algorithm became overtly racist and sexist. Without better oversight, algorithms that are supposed to help the public sector manage welfare benefits may discrimina­te against families that are in real need. Equally worrying, public authoritie­s in some countries are already using Ai-powered facial-recognitio­n technology to monitor political dissent and subject citizens to mass-surveillan­ce regimes.

Market concentrat­ion is also a major concern. AI developmen­t – and control of the underlying data – is dominated by just a few powerful players in just a few locales. Between 2013 and 2021, China and the United States accounted for 80 per cent of private AI investment globally. There is now a massive power imbalance between the private owners of these technologi­es and the rest of us.

But AI is being boosted by massive public investment as well. Such financing should be governed for the common good, not in the interest of the few. We need a digital

architectu­re that shares the rewards of collective value creation more equitably.

The era of light-touch self-regulation must end. When we allow market fundamenta­lism to prevail, the state and taxpayers are condemned to come to the rescue after the fact – as we have seen in the context of the 2008 financial crisis and the COVID-19 pandemic – usually at great financial cost and with long-lasting social scarring.

Worse, with AI, we do not even know if an ex-post interventi­on will be enough. As The Economist recently pointed out, AI developers themselves are often surprised by the power of their creations.

Fortunatel­y, we already know how to avert another laissez-faire-induced crisis. We need an ‘ethical by design’ AI mission that is underpinne­d by sound regulation and capable government­s working to shape this technologi­cal revolution in the common interest, rather than in shareholde­rs’ interest alone. With these pillars in place, the private sector can and will join the broader effort to make technologi­es safer and fairer.

Effective public oversight should ensure that digitalisa­tion and AI are creating opportunit­ies for public value creation. This principle is integral to UNESCO’S ‘Recommenda­tion on the Ethics of AI’, a normative framework that was adopted by 193 member states in November 2021. Moreover, key players are now taking responsibi­lity for reframing the debate, with US President Joe Biden’s administra­tion proposing an AI Bill of Rights, and the European Union developing a holistic framework for governing AI.

Still, we also must keep the public sector’s own uses of AI on a sound ethical footing. With AI supporting more and more decision-making, it is important to ensure that AI systems are not used in ways that subvert democracy or violate human rights.

We also must address the lack of investment in the public sector’s own innovative and governance capacities. COVID-19 has underscore­d the need for more dynamic public- sector capabiliti­es. Without robust terms and conditions governing public-private partnershi­ps, for example, companies can easily capture the agenda.

The problem, however, is that the outsourcin­g of public contracts has increasing­ly become a barrier to building public-sector capabiliti­es. Government­s need to be able to develop AI in ways that they are not reliant on the private sector for sensitive systems, so that they can maintain control over important products and ensure that ethical standards are upheld. Likewise, they must be able to support informatio­n sharing and interopera­ble protocols and metrics across department­s and ministries. This will all require public investment­s in government capabiliti­es, following a mission-oriented approach.

Given that so much knowledge and experience is now centred in the private sector, synergies between the public and private sectors are both inevitable and desirable. Mission orientatio­n is about picking the willing – by co-investing with partners that recognise the potential of government-led missions.

The key is to equip the state with the ability to manage how AI systems are deployed and used, rather than always playing catch-up. To share the risks and rewards of public investment, policymake­rs can attach conditions to public funding. They also can, and should, require Big Tech to be more open and transparen­t.

Our societies’ future is at stake. We must not only fix the problems and control the downside risks of AI, but also shape the direction of the digital transforma­tion and technologi­cal innovation more broadly.

At the start of a new year, there is no better time to begin laying the foundation for limitless innovation in the interest of all.

Gabriela Ramos is Assistant Director General for Social and Human Sciences at UNESCO. Mariana Mazzucato, founding director of the UCL Institute for Innovation and Public Purpose, is chair of the World Health Organizati­on’s Council on the Economics of Health for All.

 ?? ?? Gabriela Ramos
GUEST COLUMNIST
Gabriela Ramos GUEST COLUMNIST
 ?? ?? Mariana Mazzucato
GUEST COLUMNIST
Mariana Mazzucato GUEST COLUMNIST

Newspapers in English

Newspapers from Jamaica