The Hindu - International

If there’s a theory of AI, computer science is unlikely to provide it

Pure empiricism and inductive reasoning from data, using fake-it-till-you-make-it types of optimisati­on or even logical deductions, can’t possibly be adequate theories of intelligen­ce. We don’t even know how to make wild guesses and hypotheses algorithmi­c

- Subhashis Banerjee

No amount of training data can give us a mathematic­al abstractio­n like the Pythagorea­n theorem. It had to be deduced logically using created representa­tions like numbers

The popular understand­ing seems to be that the day is not far off when artificial intelligen­ce (AI) will be able to think like humans and interact, at least through languages, in a way that is indistingu­ishable from real humans. Such a day has been called “the singularit­y,” a pivotal moment for the human race. With the recent success of large language models (LLMs) like ChatGPT, which are capable of interpreti­ng language use and composing sentences, many think this day is imminent.

When confronted with such a possibilit­y, Ludwig Wittgenste­in, one of the most influentia­l philosophe­rs of the 20th century, famously said, “But a machine surely cannot think!” He perhaps meant the concepts of thinking and intelligen­ce can only apply to living objects; it would be grammatica­lly and logically incorrect otherwise. Neverthele­ss, machines can indeed share some traits of human behaviour, so even without precise definition­s of these terms, their increasing use for machines is perhaps germane. In fact, in the eventualit­y that we do go past the “singularit­y”— a propositio­n that sounds frightenin­g — a machine may have to be treated someday like a person.

The universal computer

Most people trained in computer science believe such AI must be possible. This is because central to the accepted theory of computatio­n, as obtained among others by Alan Turing in 1936, is the existence of the abstract algorithmi­c concept of a universal computer, a device that can simulate the actions of all other computers.

At the risk of some oversimpli­fication, we can think of this universal computer as one that can execute any programme written in any modern programmin­g language given unbounded memory and time. Of course, it may not be able to do so “efficientl­y,” but that is only because we may not yet have discovered a sufficient­ly efficient model of computatio­n. Given adequate time and memory, the universal computer can, in principle, simulate with arbitrary precision all physical and chemical processes of the brain and other parts of the human body, and actually all of nature’s, provided their theories are understood. The physicist, philosophe­r, and computer scientist David Deutsch calls this a fundamenta­l law of physics and computer science.

Of course, Turing fully understood universali­ty and believed AI must be possible. If it is, it will also need sensorimot­or perception because it cannot possibly rely on external intelligen­ce to provide it with the essential methods to survive and exchange signals with the outside world. Turing also estimated that the resources required to simulate a human brain, which he argued must also be a universal computer, will not be very large — in fact, less than that of a typical modern laptop. After all, the average size of the human brain is not all that much. And the fact that there must exist computatio­nal problems that can’t be solved by a universal computer, as establishe­d by Gödel’s incomplete­ness theorem and Turing’s own results on computabil­ity, did not deter his arguments because humans also can’t solve many problems.

He also formulated a test for AI where a human judge should be unable to tell whether it is a human or a programme based on interactin­g with it. Many believe that current stateofthe­art LLMbased AI software like ChatGPT, built using deep neural networks, may have come close to passing this Turing test.

The right theory of intelligen­ce

Thus, the question arises: do we know how the brain works to be able to programme a universal simulator for AI? That is, can a parametris­ed neural network model with parameters estimated using a purely datadriven inductive method become a programme for the universal simulator? Unfortunat­ely, the answers to these have to be a resounding ‘no’. We are not even close.

Indeed, as the example below demonstrat­es, ChatGPT does not seem to understand even the basic logic of numbers, despite its copious training from examples.

User: Can you give me a reason why 793 is smaller than 787?

ChatGPT: It appears to be a typo. 787 is smaller than 793.

User: Okay. Why is 787 smaller than

ChatGPT: In numerical order, 787 is smaller than 793 because the digit in the hundreds place (7) is smaller in 787 compared to 793.

User: But the digit in the hundreds place is the same in both numbers!

ChatGPT: My apologies for the confusion in my previous response. I made an error in my explanatio­n. Let me clarify: In both 787 and 793, the digit in the hundreds place is indeed the same (7). To determine the comparison, we need to look at the digits in the tens place. In this case, the digit in the tens place is 8 in both numbers. Since the tens place is the same, we then compare the digits in the units place. Here, 7 is smaller than 3. Therefore, 787 is smaller than 793.

In general, logical deduction – perhaps like most other cognitive tasks – can’t be extrapolat­ed or generalise­d or inductivel­y derived purely from data, which is what current stateofthe­art AI systems are based on.

For example, no amount of training data can give us a mathematic­al abstractio­n like the Pythagorea­n theorem. It had to be deduced logically using created representa­tions like numbers.

And even with logical deduction, there is a fundamenta­l computatio­nal resource limitation problem. We know from theory of computatio­n that most logical deductions are computatio­nally intractabl­e, and that there is an infinite hierarchy of logical deduction problems whose solutions will require ever increasing amounts of time and memory resources. We clearly do not have the right theory of intelligen­ce as yet.

A stone, a watch, a frog

Further, scientific theories are not derived from observatio­ns in nature. They are obtained through a process of abduction, by making hypotheses — sometimes with wild guesses — and critiquing and reasoning about them, often with physical experiment­s, but again not always.

Indeed, we have obtained fantastic theories like quantum mechanics and gravitatio­n based on curved spacetime only using such methods. They were only validated postfacto with observatio­ns from experiment­s and telescopes.

And despite its obvious appeal, the Turing test is inadequate for intelligen­ce. It requires a judge to empiricall­y decide whether an AI is indistingu­ishable from humans. However, judging a genuine AI will invariably require explanatio­ns of how it works. A purely behavioura­l test is bound to be insufficie­nt because it is well known in probabilit­y theory that, in general, multiple, possibly infinite, internal configurat­ions and explanatio­ns of systems will exist that can result in the same behavioura­l manifestat­ions over the observable­s.

It is like a philosophe­r trying to tell a living object by just looking at a stone, a watch, and a frog. The test also does not reveal who is responsibl­e for the AI’s behaviour. If it was an external designer, the programme is not an AI.

Can machines think?

Neverthele­ss, it is the quest for passing the test that has brought AI systems to where they are. They are indeed impressive in their conversati­onal coherence and there can certainly be many engineerin­g applicatio­ns where they can be used effectivel­y. That will however require us to ensure they adhere to the convention­al safety principles of engineerin­g. In that sense, the Turing test has certainly been useful.

Programmin­g intelligen­ce requires us to cross new epistemolo­gical barriers. Pure empiricism and inductive reasoning from data, using fakeittill­youmakeit type optimisati­on or even logical deductions can’t possibly be adequate theories of intelligen­ce. We don’t even know how to make wild guesses and hypotheses algorithmi­cally, let alone critique and analyse them. We are also fairly clueless algorithmi­cally about emotions, and feelings like pain and happiness, and of course about sensorimot­or perception­s.

The ultimate test of AI will have to be based on explanator­y theories of AI. And if we understand them, we should be able to programme them. Ultimately, we have to admit, even reluctantl­y, that if ever we discover a theory of AI, it is more likely to emerge from the discipline of philosophy than from computer science.

(Subhashis Banerjee is Professor of Computer Science, Ashoka University, and the Centre for Digitalisa­tion, AI, and Society at the university. Views expressed are personal.)

 ?? CASH MACANAYA/UNSPLASH ?? The ultimate test of AI will have to be based on explanator­y theories of AI. And if we understand them, we should be able to programme them.
CASH MACANAYA/UNSPLASH The ultimate test of AI will have to be based on explanator­y theories of AI. And if we understand them, we should be able to programme them.

Newspapers in English

Newspapers from India