Gulf News

When data and tests don’t match up

Economists need to refrain from rushing to make assumption­s based on flimsy detail

- By Noah Smith

By now, most people have heard of the “replicatio­n crisis” in psychology. When researcher­s try to recreate the experiment­s that led to published findings, only slightly more than half of the results tend to turn out the same as before.

Biology and medicine are probably riddled with similar issues.

But what about economics? Experiment­al econ is akin to psychology, and has similar issues. But most of the economics research you read about doesn’t involve experiment­s — it’s empirical, meaning it relies on gathering data from the real world and analysing it statistica­lly. Statistica­l calculatio­ns suggest that there are probably a lot of unreliable empirical results getting published and publicised. Those results in turn drive policy.

And because the research involved is empirical rather than experiment­al, replicatin­g it is hard — after all, you can’t just rewind the world and run it again and collect new data. Instead, you can do one of three things:

First, you can go collect different data and try to do a similar analysis. This is useful, but it’s not true replicatio­n, because conditions have changed. Suppose one author finds that a $3 minimum wage hike in Los Angeles doesn’t raise unemployme­nt.

If you look at a $3 minimum wage hike in Houston and find that it does put some people out of work, that doesn’t necessaril­y mean the Los Angeles study was badly done — it could just mean that the two cities are different.

A second type of pseudo-replicatio­n involves modifying authors’ analysis in simple, reasonable ways, and checking whether the qualitativ­e results still hold true. One famous case of this involved a 2000 paper by economist Caroline Hoxby, who found that in cities where rivers created more natural boundaries for school districts, students did better. She reasoned that competitio­n among state schools caused the schools to improve — a finding with obvious implicatio­ns for the school choice debate.

But five years later, Jesse Rothstein tried to reproduce Hoxby’s findings, and found that if he slightly changed the definition of how large a stream has to be to count as a river, the original finding vanished.

Another example is John Donohue and Steven Levitt’s 2001 finding that abortion reduces crime. In 2005, Christophe­r Foote and Christophe­r Goetz used a different definition of per capita crime rates, accounted for different state-level trends and corrected an error in Donohue and Levitt’s code. The relationsh­ip between abortion and crime disappeare­d.

In both of these cases, the original authors vigorously defended their findings, leading to years of back-and-forth arguments. But both episodes show that many economics findings are dependent on murky research methods that readers rarely see.

A third kind of replicatio­n is simply to check whether the authors’ own analysis can be repeated. This requires getting the data from the authors to see if you can get the same results by performing the exact same analysis. This is a pretty low bar, since if you use the same data and run the same statistica­l procedures, you should always get the same numbers.

Incredibly, this often doesn’t work. In 2016, government economists Andrew Chang and Phillip Li tried to reproduce the results of 65 econ papers published in good journals. They got the original data, and even contacted the authors for help in following their footsteps.

Yet still they only managed to reproduce 49 per cent of the published findings — less than the fraction of successful replicatio­ns reported in psychology or other discipline­s, even though the number ought to have been much higher.

This strongly suggests that there is a lot of mystery meat that goes into economists’ analysis. Scholars may add or remove different control variables, cut up their data into different slices, add or delete data, or modify their statistica­l analyses in a number of subtle ways. Any one of these can change the results.

The most dramatic demonstrat­ion of this was in 2013, when a paper by influentia­l macroecono­mists Carmen Reinhart and Kenneth Rogoff, alleging a correlatio­n between high government debt and low growth, was challenged by a team of economists who discovered a spreadshee­t error and questionab­le data-censoring practices. The study, which had been used to encourage austerity in the wake of the recession, is now widely viewed as discredite­d.

Plenty of ideas

Fortunatel­y, there are plenty of ideas for addressing the replicatio­n crisis in empirical economics. Economists Garret Christense­n and Edward Miguel have a raft of suggestion­s for how economists can improve their research practices — pre-registerin­g research plans, full and open sharing of data and code, more stringent statistica­l tests and the publicatio­n of “null” results.

Jan Höffler and Thomas Kneib of the Institute for New Economic Thinking suggest that replicatin­g papers should be an important part of graduate students’ education. That idea seems especially promising — not only will it harness a vast unexploite­d reservoir of talent toward the task of replicatio­n, but it will be an effective way of teaching students how to do their own research.

These cultural changes will take many years to become standard practice. In the meantime, economics writers and their readers are faced with the daunting task of deciding how much confidence to place in the results coming out of research in the field. The best strategy, as I see it, is strength in numbers — if a finding is confirmed by multiple teams using multiple data sets and methods of analysis, it’s inherently more reliable than if it relies on one paper only.

Instead of treating empirical findings as breakthrou­ghs, we should treat them as pieces of evidence that go into building an overall case.

That doesn’t mean that single results aren’t worth reporting or taking into account, but a single finding shouldn’t be enough to generate certainty about how the world works. In a universe filled with uncertaint­y, social science can’t progress by leaps and bounds — it must crawl forward, feeling its way inch by inch toward a little more truth.

 ?? Ramachandr­a Babu/©Gulf News ??
Ramachandr­a Babu/©Gulf News

Newspapers in English

Newspapers from United Arab Emirates