Arab News

The dangers of relying on data

- DIANE COYLE For full version, log on to www.arabnews.com/opinion

Friends of mine who work in the arts and humanities have started doing something unusual, at least for them: Poring over data. This is due to the pandemic, of course. Every day, they check coronaviru­s case numbers, how slowly or quickly the R number is declining, and how many people in our area got vaccinated the day before. Meanwhile, social media sites are full of claims and countercla­ims about all manner of other data. Is global poverty declining or increasing? What is the real level of US unemployme­nt? The scrutiny, sometimes leading to tetchy arguments, results from people’s desire to cite — or challenge — the authority of data to support their position or worldview.

But, in other areas where data is used, there is remarkably little focus on its reliabilit­y or interpreta­tion. One striking example concerns the “CAPTCHA” tests designed to protect websites against bots, which ask you to prove your humanity by identifyin­g images containing common features such as boats, bicycles or traffic lights. If your choice — even if correct — differs from that of the machine system using your selection to train an image-recognitio­n algorithm, you will be deemed inhuman.

In this example, the machine’s error is obvious. But, in other cases, it may not be possible to identify what conclusion­s either machine-learning systems or human analysts are drawing when they put more weight on data than the data can bear.

Many policymake­rs think artificial intelligen­ce offers scope for greater cost-effectiven­ess and better policy outcomes. But before we entrust more decisions to data-based systems, we must be clear about the limitation­s of the data.

The data we use shapes our view of a complex, changing world.

But data represents reality from a particular perspectiv­e. Data of the kind deployed in policy debates is rarely completely unanchored from the world it describes, but the lens it provides can be sharp or blurry — and there is no escaping the perspectiv­e it offers.

The current hunger for databased certainty is becoming dangerous as we increasing­ly rely on technocrat­ic decision procedures — including machinelea­rning systems — for policymaki­ng in areas such as criminal justice, policing and welfare. Democracie­s often rely on constructi­ve ambiguity to reconcile conflictin­g interests, such as those regarding the distributi­on of returns to an asset or to address the question of whether law-enforcemen­t authoritie­s should err on the side of imprisonin­g innocent people or letting criminals walk free. Claims to data-based authority minimize or eliminate the scope of ambiguity, with potentiall­y significan­t consequenc­es.

I am all in favor of more and better data, which has been essential to government­s’ efforts to manage the pandemic. But the more we use data to make decisions, the more sensitive we must be to the fact that data paints an expert’s or machine’s-eye view, based on categories devised by someone who is themselves a player in society’s status game. Otherwise, we will end up with decision processes just like those rogue CAPTCHA tests — insisting that a boat is a bicycle and leaving other people with no choice but to agree.

Diane Coyle, Professor of Public Policy at the University of Cambridge, is the author, most recently, of “Markets, State, and People: Economics for Public Policy.”

©Project Syndicate

 ??  ??

Newspapers in English

Newspapers from Saudi Arabia