National Post (National Edition)

CONSIDER THIS A WARNING

‘CORRELATIO­N IS NOT CAUSATION’ IS NOT A MAGIC PHRASE THAT WINS EVERY ARGUMENT

- STEPHEN GORDON Stephen Gordon is a professor of economics at Université Laval.

If you’ve taken a course in statistics, you’ve probably been warned against mistaking correlatio­n for causality, and it’s good advice. But this is one of those cases where a little knowledge can be a dangerous thing: “correlatio­n is not causality” is not an incantatio­n that gives you the power to think of yourself as having beliefs based on data while dismissing out of hand any evidence that challenges your previously-held opinions.

To be sure, correlatio­n and causality are different things, and data can never prove causality. Even in a (relatively) simple context such as a billiards table, causal relations cannot be observed directly. Causality is part of the theoretica­l framework that underlies our understand­ing of how billiard balls move, not something we can measure. As the 18th-century philosophe­r David Hume observed in An Enquiry Concerning Human Understand­ing, we don’t actually observe the transfer of momentum from one billiard ball to the next.

It’s easy to come up with a causal explanatio­n for what we see. Too easy, in fact: there are an infinite number of theories that can potentiall­y explain a finite set of facts. Isaac Newton’s explanatio­n for what happens after one billiard ball hits another is not the only one: there’s always the possibilit­y that everything is being manipulate­d by the Flying Spaghetti Monster’s invisible tentacles. Or that everything we perceive is part of the Matrix or that we’re just a dream someone is having. The possibilit­ies are endless, and not all of them have been made into films yet. The way to avoid going further down that rabbit-hole is to limit ourselves to theories that make prediction­s that might be wrong, or falsifiabl­e.

Of course, no theory perfectly forecasts every outcome: measuremen­t errors and confoundin­g factors (friction, weather, politics, etc.) impose limits on how precise a prediction can be. This is where probabilit­y and statistics come in: theories that produce smaller errors are given more credence. The more informatio­n there is — more data, or data that clearly favour one side over the other — then the easier it is to decide which theory is in line with the data.

The real problem occurs when the same correlatio­n is consistent with more than one theory: much of the time spent on empirical projects consists of trying to make sure that the effect we’re measuring is the effect we’re interested in. If prices and quantities are determined simultaneo­usly by the interactio­ns of supply and demand, then there’s no way of telling if a correlatio­n between prices and quantities reflects a supply effect, a demand effect, or a combinatio­n of both. This “identifica­tion problem” is pervasive in economics.

This is why people prefer working with data generated by controlled experiment­s when they can. A properly designed experiment will ensure that the variation in the (hypothesiz­ed) causal factor is purely random, and not subject to feedback from the effect you’re trying to identify. Unfortunat­ely, experiment­al data are hard to come by in economics. Some economists have taken to performing their own experiment­s — using laboratory techniques similar to those used in psychology — but we’re not yet at the point where lessons learned in the lab are routinely applied to policy problems. Randomized controlled trials (RCT) — in which people are chosen at random to receive a “treatment” — are still rare, and are usually not feasible.

“Natural experiment­s” occur from time to time, when what amounts to a random change in the economic environmen­t — a policy change, a natural disaster — affects only part of the population. In this case, the difference­s in outcomes in the two subgroups can be potentiall­y explained by the fact that one received the treatment, and the other did not.

But for the most part, economic data are non-experiment­al (or observatio­nal) — just as they are for the other social sciences and for some natural sciences such as fields like astronomy and climatolog­y. In this case, you have to make explicit assumption­s about how the world works — that is, set out a model specifying the theory of causality — and compare its prediction­s to the data. This modelling exercise is at the heart of all empirical studies that try to extract causality from correlatio­n, and you can’t cheat by saying that everything depends on everything else. If you don’t have a controlled experiment to close off some causal relations, you have to close them off by assumption. Without these “identifyin­g” restrictio­ns, you can’t make the connection between correlatio­n and causality.

These restrictio­ns are definitely up for debate: you can’t appeal to the data to test if an identifica­tion strategy is valid. These assumption­s have to stand on their own, on the basis of their intuitive plausibili­ty and/or other empirical work. For example, a popular identifica­tion restrictio­n that says that technical change and trade liberaliza­tion affects labour demand, but not labour supply. This may not be literally true — it’s easy enough to come up with some sort of multi-step counterfac­tual — but it’s plausible enough for empirical work.

It’s fair game to question identifica­tion assumption­s, but the real challenge is to come up with an even more plausible set of restrictio­ns. Unless you have a more plausible identifica­tion strategy for interpreti­ng the results, saying “correlatio­n is not causation” to deflect a conclusion you don’t like is the equivalent of putting your fingers in your ears and yelling “La la la, I can’t hear you.”

This happens all the time. My favourite recent example was a recent study linking the introducti­on of universal daycare in Quebec to a decline in children’s non-cognitive skills. This resulted in a wave of indignant denunciati­ons of its “flawed” methodolog­y, but of course the study’s real sin was in producing a conclusion that displeased many people. Meanwhile, studies that use the same identifica­tion assumption­s and find that universal daycare increased female labour force participat­ion rates in Quebec were met with nods of approval. When the same methodolog­y produces two sets of results and you reject only one, that’s confirmati­on bias at work, not analysis.

 ??  ??

Newspapers in English

Newspapers from Canada