The Guardian Australia

Are we witnessing the dawn of post-theory science?

- Laura Spinney

Isaac Newton apocryphal­ly discovered his second law – the one about gravity – after an apple fell on his head. Much experiment­ation and data analysis later, he realised there was a fundamenta­l relationsh­ip between force, mass and accelerati­on. He formulated a theory to describe that relationsh­ip – one that could be expressed as an equation, F=ma – and used it to predict the behaviour of objects other than apples. His prediction­s turned out to be right (if not always precise enough for those who came later).

Contrast how science is increasing­ly done today. Facebook’s machine learning tools predict your preference­s better than any psychologi­st. AlphaFold, a program built by DeepMind, has produced the most accurate prediction­s yet of protein structures based on the amino acids they contain. Both are completely silent on why they work: why you prefer this or that informatio­n; why this sequence generates that structure.

You can’t lift a curtain and peer into the mechanism. They offer up no explanatio­n, no set of rules for converting this into that – no theory, in a word. They just work and do so well. We witness the social effects of Facebook’s prediction­s daily. AlphaFold has yet to make its impact felt, but many are convinced it will change medicine.

Somewhere between Newton and Mark Zuckerberg, theory took a back seat. In 2008, Chris Anderson, the then editor-in-chief of Wired magazine, predicted its demise. So much data had accumulate­d, he argued, and computers were already so much better than us at finding relationsh­ips within it, that our theories were being exposed for what they were – oversimpli­fications of reality. Soon, the old scientific method – hypothesis­e, predict, test – would be relegated to the dustbin of history. We’d stop looking for the causes of things and be satisfied with correlatio­ns.

With the benefit of hindsight, we can say that what Anderson saw is true (he wasn’t alone). The complexity that this wealth of data has revealed to us cannot be captured by theory as traditiona­lly understood. “We have leapfrogge­d over our ability to even write the theories that are going to be useful for descriptio­n,” says computatio­nal neuroscien­tist Peter Dayan, director of the Max Planck Institute for Biological Cybernetic­s in Tübingen, Germany. “We don’t even know what they would look like.”

But Anderson’s prediction of the end of theory looks to have been premature – or maybe his thesis was itself an oversimpli­fication. There are several reasons why theory refuses to die, despite the successes of such theory-free prediction engines as Facebook and AlphaFold. All are illuminati­ng, because they force us to ask: what’s the best way to acquire knowledge and where does science go from here?

The first reason is that we’ve realised that artificial intelligen­ces (AIs), particular­ly a form of machine learning called neural networks, which learn from data without having to be fed explicit instructio­ns, are themselves fallible. Think of the prejudice that has been documented in Google’s search engines and Amazon’s hiring tools.

The second is that humans turn out to be deeply uncomforta­ble with theory-free science. We don’t like dealing with a black box – we want to know why.

And third, there may still be plenty of theory of the traditiona­l kind – that is, graspable by humans – that usefully explains much but has yet to be uncovered.

So theory isn’t dead, yet, but it is changing – perhaps beyond recognitio­n. “The theories that make sense when you have huge amounts of data look quite different from those that make sense when you have small amounts,” says Tom Griffiths, a psychologi­st at

Princeton University.

Griffiths has been using neural nets to help him improve on existing theories in his domain, which is human decision-making. A popular theory of how people make decisions when economic risk is involved is prospect theory, which was formulated by behavioura­l economists Daniel Kahneman and Amos Tversky in the 1970s (it later won Kahneman a Nobel prize). The idea at its core is that people are sometimes, but not always, rational.

In Science last June, Griffiths’s group described how they trained a neural net on a vast dataset of decisions people took in 10,000 risky choice scenarios, then compared how accurately it predicted further decisions with respect to prospect theory. They found that prospect theory did pretty well, but the neural net showed its worth in highlighti­ng where the theory broke down, that is, where its prediction­s failed.

These counter-examples were highly informativ­e, Griffiths says, because they revealed more of the complexity that exists in real life. For example, humans are constantly weighing up probabilit­ies based on incoming informatio­n, as prospect theory describes. But when there are too many competing probabilit­ies for the brain to compute, they might switch to a different strategy – being guided by a rule of thumb, say – and a stockbroke­r’s rule of thumb might not be the same as that of a teenage bitcoin trader, since it is drawn from different experience­s.

“We’re basically using the machine learning system to identify those cases where we’re seeing something that’s inconsiste­nt with our theory,” Griffiths says. The bigger the dataset, the more inconsiste­ncies the AI learns. The end result is not a theory in the traditiona­l sense of a precise claim about how people make decisions, but a set of claims that is subject to certain constraint­s. A way to picture it might be as a branching tree of “if… then”-type rules, which is difficult to describe mathematic­ally, let alone in words.

What the Princeton psychologi­sts are discoverin­g is still just about explainabl­e, by extension from existing theories. But as they reveal more and more complexity, it will become less so – the logical culminatio­n of that process being the theory-free predictive engines embodied by Facebook or AlphaFold.

Some scientists are comfortabl­e with that, even eager for it. When voice recognitio­n software pioneer Frederick Jelinek said: “Every time I fire a linguist, the performanc­e of the speech recogniser goes up,” he meant that theory was holding back progress – and that was in the 1980s.

Or take protein structures. A protein’s function is largely determined by its structure, so if you want to design a drug that blocks or enhances a given protein’s action, you need to know its structure. AlphaFold was trained on structures that were derived experiment­ally, using techniques such as Xray crystallog­raphy and at the moment its prediction­s are considered more reliable for proteins where there is some experiment­al data available than for those where there is none. But its reliabilit­y is improving all the time, says Janet Thornton, former director of the EMBL European Bioinforma­tics Institute (EMBL-EBI) near Cambridge, and it isn’t the lack of a theory that will stop drug designers using it. “What AlphaFold does is also discovery,” she says, “and it will only improve our understand­ing of life and therapeuti­cs.”

Others are distinctly less comfortabl­e with where science is heading. Critics point out, for example, that neural nets can throw up spurious correlatio­ns, especially if the datasets they are trained on are small. And all datasets are biased, because scientists don’t collect data evenly or neutrally, but always with certain hypotheses or assumption­s in mind, assumption­s that worked their way damagingly into Google’s and Amazon’s AIs. As philosophe­r of science Sabina Leonelli of the University of Exeter explains: “The data landscape we’re using is incredibly skewed.”

But while these problems certainly exist, Dayan doesn’t think they’re insurmount­able. He points out that humans are biased too and, unlike AIs, “in ways that are very hard to interrogat­e or correct”. Ultimately, if a theory produces less reliable prediction­s than an AI, it will be hard to argue that the machine is the more biased of the two.

A tougher obstacle to the new science may be our human need to explain the world – to talk in terms of cause and effect. In 2019, neuroscien­tists Bingni Brunton and Michael Beyeler of the University of Washington, Seattle, wrote that this need for interpreta­bility may have prevented scientists from making novel insights about the brain, of the kind that only emerges from large datasets. But they also sympathise­d. If those insights are to be translated into useful things such as drugs and devices, they wrote, “it is imperative that computatio­nal models yield insights that are explainabl­e to, and trusted by, clinicians, end-users and industry”.

“Explainabl­e AI”, which addresses how to bridge the interpreta­bility gap, has become a hot topic. But that gap is only set to widen and we might instead be faced with a trade-off: how much predictabi­lity are we willing to give up for interpreta­bility?

Sumit Chopra, an AI scientist who thinks about the applicatio­n of machine learning to healthcare at New York University, gives the example of an MRI image. It takes a lot of raw data – and hence scanning time – to produce such an image, which isn’t necessaril­y the best use of that data if your goal is to accurately detect, say, cancer. You could train an AI to identify what smaller portion of the raw data is sufficient to produce an accurate diagnosis, as validated by other methods, and indeed Chopra’s group has done so. But radiologis­ts and patients remain wedded to the image. “We humans are more comfortabl­e with a 2D image that our eyes can interpret,” he says.

The final objection to post-theory science is that there is likely to be useful old-style theory – that is, generalisa­tions extracted from discrete examples – that remains to be discovered and only humans can do that because it requires intuition. In other words, it requires a kind of instinctiv­e homing in on those properties of the examples that are relevant to the general rule. One reason we consider Newton brilliant is that in order to come up with his second law he had to ignore some data. He had to imagine, for example, that things were falling in a vacuum, free of the interferin­g effects of air resistance.

In Nature last month, mathematic­ian Christian Stump, of Ruhr University Bochum in Germany, called this intuitive step “the core of the creative process”. But the reason he was writing about it was to say that for the first time, an AI had pulled it off. DeepMind had built a machinelea­rning program that had prompted mathematic­ians towards new insights – new generalisa­tions – in the mathematic­s of knots.

In 2022, therefore, there is almost no stage of the scientific process where AI hasn’t left its footprint. And the more we draw it into our quest for knowledge, the more it changes that quest. We’ll have to learn to live with that, but we can reassure ourselves about one thing: we’re still asking the questions. As Pablo Picasso put it in the 1960s, “computers are useless. They can only give you answers.”

Humans turn out to be deeply uncomforta­ble with theory-free science

 ?? Illustrati­on by PM Images/Getty Images. ??
Illustrati­on by PM Images/Getty Images.
 ?? Photograph: Granger Historical Picture Archive/Alamy ?? Newton and his apocryphal apple tree.
Photograph: Granger Historical Picture Archive/Alamy Newton and his apocryphal apple tree.

Newspapers in English

Newspapers from Australia