IEEE 9

 

Home
Up

Copyright © 2005, Institute of Electrical and Electronics Engineers, Inc. All rights reserved. This article was published in IEEE Intelligent Systems magazine

This article is also available in PDF format here.

Moving AI out of its infancy: Changing our Preconceptions

Steve Grand

The other day it was my turn to be asked stupid questions about the movie ‘I, Robot’. ‘Do you think it’s about time we started incorporating Asimov’s three laws into real robots?’ a journalist asked. My reply was that Asimov’s laws are about as relevant to real robotics as leechcraft is to modern medicine. And yes, before anyone else writes me smug emails, I know that leeches are very useful in modern medicine, but what I actually said was ‘leechcraft’. Leeches may be useful, but the paradigm of thought that originally led to their use is a ridiculous anachronism.

The same is true for Asimov’s laws. Of course, he invented them as a plot device, useful precisely because of their many paradoxes and faults. Nevertheless, they do belong to a bygone age: a world created by Boole and Babbage, in which people seriously thought that intelligence was some form of logical calculus. At the time, it seemed reasonable that a robot could be given explicit rules for every aspect of its behaviour. It could be ‘programmed’ to recognise when it was in danger of breaking any laws, and could be guaranteed not to ‘break its programming’ by doing anything wrong. Within a few years of Asimov’s stories the digital computer had been invented and people seriously set to work trying to program in such rules for intelligence.

How quaint! Luckily nobody believes this nowadays. Or do they? Explicit symbolic logic has faded from prominence, but the close coupling between AI and the digital computer, and between thought and the stepwise algorithm, seem about as strong and unquestioned as ever.

Of course there’s connectionism, but this too is mired in false assumptions that date back a very long way. And it seems to have dragged neuroscience down with it, to the extent that we seem unable to think about real brains without at some level resorting to models which owe rather too much of their inspiration to the three-layer perceptron. Brains really aren’t like that.

Traditional AI has excelled at solving certain kinds of problem: it can build machines that play chess, but not ones that can pick up the pieces when they fall over. It can create machines that read text but not ones that can recognise objects from arbitrary angles. It can make systems that learn, but they don’t do it in any generally applicable way. This is not a criticism – traditional artificial intelligence mostly follows Minsky’s dictum that AI is about making machines do what humans use intelligence to do, and often this doesn’t actually require the machines to show any intelligence at all. But for many tasks, especially in robotics, the ability to see, learn and perform complex motor actions is a prerequisite that the traditional approach has utterly failed to fulfil. If robots are any kind of a threat to humanity it’s only because they tend to be heavy and fall over a lot. Even if a machine could contemplate murder it wouldn’t be able to pick up the knife or locate the victim.

The New AI fares a little better at solving some of these supposedly lower-level tasks. But where good old-fashioned AI was inspired by the logical thought processes of advanced mathematicians, New AI is inspired by the nervous systems of the very simplest invertebrates (and no, that’s not the same thing). The snag is that these extreme bottom-up and top-down approaches don’t meet in the middle. There’s a huge gulf, precisely where the interesting behaviour lies. Neither approach tells us much about how to make machines that can perceive or make complex movements in the way that even the most primitive mammals can, yet these are the very competencies that robots so desperately need. And despite what some people seem to assume, you can’t simply combine techniques from both approaches. A human being is not an ant with a natural language interface.

AI currently stands in relation to real intelligence much as alchemy once did to chemistry. Without alchemists we would never have developed chemistry in the first place, so I mean no insult by this. But until the discovery of the periodic table everyone was essentially stabbing in the dark. Fundamentally, what we’ve learned over the past fifty years is a lot about how not to build intelligent machines, but we still haven’t made the critical breakthrough. There is one class of machine that we know for sure can solve all these problems of perception and complex action: the mammalian brain. But we simply don’t know what its fundamental operating principles are (although I’m quite certain it has some). We have no periodic table of neural function to help us see the underlying logic. Turing machines and neural networks were hopeful candidates for a theory of intelligence but they simply don’t cut the mustard.

Because of the way science works, we have a tendency to follow bandwagons, for the most part making only incremental improvements to ideas that already exist. But a beautifully polished and optimised bad idea is still a bad idea. The digital computer was inspired by one abstract view of the process of thought, but it turns out to have been the wrong one as far as general intelligence is concerned. Connectionists tried to pay more attention to the neural hardware, but they did so within a paradigm drawn from electronic circuitry, causing them to make rash assumptions about the roles of nerve cells and synapses. Neither approach has worked, so we should abandon these paradigms and look for other models. What we need are some new and radical ideas at the most fundamental level.

Edward de Bono, the champion of lateral thinking, has a technique which he calls Po. It involves making deliberately provocative statements (‘the best place to sell ice cream is the North Pole’), in order to shake us out of our preconceptions and find new paths. Suppose de Bono were to take up AI. What kinds of Po statement might he make?

I really couldn’t say, but the following are some of the deliberate provocations that stimulate my own research. To my mind they are nowhere near as radical as marketing ice cream to Eskimos, although some may think so. Nevertheless, I think they are sufficiently misaligned with established wisdom to suggest some interesting new directions. I ask you to treat them in the spirit of Po: not as something to criticise but as ideas to run with, just to see where they might lead.

Po 1: The purpose of the brain is to compensate for all the time it takes for the nerve signals to travel to it and back again.

Modern AI has a terrible tendency to treat intelligence as a reactive, even a passive process. AI Nouveau has a particular mistrust of internal models and top-down mechanisms, arising from a justifiable disenchantment with symbolic representation. But for real animals larger than a pinhead, prediction is an absolutely essential part of all intelligent behaviour, and reactive solutions simply won’t do. Turning your eyes towards the point where a fast-moving object was when the light from it hit your retina will cause you to miss it by miles. Equally, you have to start slowing a heavy limb long before it reaches its target or it will overshoot. Worse still, waiting until a lion has actually eaten you is not the best time to think about a response.

This principle applies universally. Indeed, perhaps the intelligence of an animal is by definition proportional to its degree of predictive power. Relatively primitive animals live for the moment, but even they still have to be able to extrapolate the movements of their prey, or predict the likely response of a social rival. Humans can form predictions extending many years into the future, or many steps into a tree of possibilities. Intelligence is all about prediction. Brains are fundamentally there to ask ‘what next?’, and in some animals, ‘what if?’

Both questions imply a mental model of the world. Not a symbolic model, nor even an explicit one, but a model nonetheless. Without some means of fast-forwarding the present it is impossible to anticipate the future, especially when that future is highly conditional. Extreme reactivists may disagree with this, but finding reactive alternatives often requires absurd contortions and flies in the face of a large body of evidence. Thinking about where and how the brain could develop, store and utilise such a model or models can be a remarkably productive exercise, when freed from any historical baggage, and can suggest unifying principles that extend from simple reflexes right through to conscious imagery.

Po 2: Brains don’t make decisions

They simply try to reduce the tension between how things are and how we expect or would like them to be.

AI has shown a lamentable tendency to slide from reasonable observations into over-stylized, formal solutions. Of course brains make decisions, but it doesn’t follow that there is an explicit decision-making mechanism in the brain in the sense used in action selection networks, etc. Much of what the brain does requires analogue, ‘left hand down a bit’ kinds of responses, and yet so many AI techniques (and quite a lot of behaviourist psychology) presume that decisions are all-or-nothing, discrete choices.

The brain must contain an anticipation of the future state of the world in order to act in good time, and to construct this it must also have a representation of the present (or more strictly the recent past) produced by the senses. At any one moment, therefore, the brain contains two complex vectors, one pattern of nerve activity representing how things are now, and the other representing how things might be soon. It makes good neurological and psychological sense to assume that these two state vectors map onto the same brain territory. If so, then the two can be directly compared, point for point, and the comparison can yield useful consequences. My present research started out with the assumption that brains are in essence arrays of servomotors – each comparing one pair of ‘values’ from the two state vectors and producing an output designed to reduce the difference between them, either by driving muscles or by becoming the ‘intention’ value for another servo in the network.

When you think about it, an intention is actually a kind of prediction. It’s a prediction about how the world will look if things go as planned. Perception and intention are thus closely related. The difference between how things are now and how we would like them to be tells us something useful about what we need to do. Our goal is to bring the state of the world into line with our ‘prediction’. This is servo action. But equally, sometimes we need to bring our predictions back into line with reality instead. This is what happens when we update our ‘beliefs’ in the light of new information. Beliefs, hypotheses, expectation, attention, plans and intention all become the same thing when seen in this light.

Po 3: Brains perform coordinate transforms.

If brains are networks of servos, each servo must operate within a particular coordinate space (for example retinotopic, somatotopic or tonotopic) and the links between them must therefore carry out a conversion from one space to another. If something that we see is to have an effect on what we do, then it follows that information mapped in retinal coordinates eventually has to produce changes in the brain that are mapped in muscle coordinates.

The more I think about this principle, the more I find that other kinds of mental process can be described as coordinate transforms too. Even object recognition and abstract symbol manipulation. One especially interesting example may explain how it is that we can recognise shapes by sight or by touch, regardless of their scale, rotation and position on the visual field or the skin. Such invariance is easily the most striking and challenging aspect of perception (and indeed motor action). Without the ability to replicate this property we have no hope of making robots that can see like animals do.

So is there a coordinate frame in which a banana, for sake of argument, looks exactly the same shape, regardless of its position, orientation and scale? Yes there is. It’s a rather abstract frame, but try this: Imagine the image of a banana falling on your retina. Now mentally project yourself until you are inside the banana, looking outwards. What you have just performed is a conversion from eye-centred coordinates into banana-centred coordinates. From inside, the banana remains exactly the same shape, regardless of its location and orientation with respect to your original viewpoint. So if brains can perform on-the-fly transforms from egocentric to object-centred coordinate space, they have the means to develop visual invariance.

Exactly how this might happen is an unsolved problem, but it’s something I’ve been working on with a modicum of success – enough to suggest that it is a meaningful idea. Significantly, if a general mechanism can be found, it will bring the two key visual data streams (the ‘where’ pathway of the parietal lobes, and the ‘what’ pathway of the temporal lobes) into a common level of explanation.

Po 4: Nervous tissue is a new state of matter

The more I think about concepts like servos and coordinate spaces, the more irrelevant the traditional view of the neuron seems to become. The stereotypical neural network is a sparsely connected, discrete signalling system, but real neurons are nothing of the kind. They are so densely interconnected and leaky that an unbiased appraisal of the facts would suggest that nervous tissue is more like a wobbly jelly than a printed circuit board. Signals spread out very rapidly and large-scale phenomena such as waves build up on the neural surface. Yet at the same time, neurons do have the ability to make changes to signal propagation on a very fine scale indeed. It seems to me that nervous tissue is a substance with some of the properties of a discrete network of wires and some of a continuous solid.

On such a medium, patterns of nerve activity need to be interpreted very differently than in conventional models based on very small networks. Perhaps it isn’t the neurons that perform the computations in the brain at all. Perhaps they provide the surface upon which the patterns of nerve activity perform computations [see the work of Steven Lehar for one interpretation of such second-order computation]. You wouldn’t learn anything about the behaviour of the Niagara Falls by studying a handful of water molecules, so it seems ridiculous to try and mimic nervous systems using simulations built from sixteen neurons. If the scale of activity is as large as I suspect it is, we simply won’t be able to see the wood for all the trees.

Po 5: The more complex a robot is, the easier it is to make progress

A similar argument about scale applies to robots, and to the segregation of disciplines within AI. Toy environments are often far too stylised and reduced to capture the essential features of a problem, and with robots that’s especially true. How intelligent would you have become if you’d been born equipped with only two wheels and a handful of bump sensors?

But there are other reasons to think big too. Hearing, seeing, planning and moving seem on the surface to be radically different problems, and yet the brain tissues involved in each of these processes are fundamentally similar. Motor cortex looks slightly different from primary visual cortex but the essential architecture is the same and most of the differences are likely to be a consequence of adaptation. So if the same brain architecture can perform all these different tasks, there must be a level of description at which they are the same task.

Moreover, so much of development and learning is multi-modal. How can we learn to see depth unless we have the ability to reach out and touch things to confirm how far away they are? How can we learn to reach out and touch things unless we can see their depth? Learning is a process of integration, correlation and confirmation between all the senses and motor systems at once, so we need to study them together.

So, to help me develop my ‘Po’ ideas about possible new neural computing architectures I decided to build as complex a robot as my limited resources would allow. Her name is Lucy, and she’s ostensibly a robot orang-utan, although the similarity is minimal because I’m not much of an artist. Nevertheless I gave her a face and refer to her as ‘her’ rather than ‘it’ in order to remind me that I’m trying to make a somebody, not a something – a complete integrated organism. Physically, she has vision, hearing, proprioception, a virtual model of the vocal tract and enough degrees of freedom to make movement a challenge. My goal is to build Lucy a brain from scratch, facing up to the same problems that nature must have faced, armed with the same tools and equipment (as far as I can manage). What I’m trying to find is a common level of description that marries all the apparently disparate tasks which brains carry out. I’m essentially on the look-out for a proto-machine: a generalised neural architecture that can spontaneously self-organise into a variety of specialised machines, driven only by the nature of the signals supplied to it.

So far, Lucy’s only party trick is that she has learned how to point at bananas. I can hold up an apple and a banana and she can (or rather could, before I started building Lucy MkII) point at the banana. Impressive, huh? And all that this feat requires is a neural network composed of around 50,000 complex neurons. Not the most efficient way to recognise a banana, when ‘point at the yellow bit’ would suffice. But I don’t care because I’m only using digital computers as an interim solution. My aim is to find radically new kinds of computing device that work more like I think brains do.

Lucy’s virtual brain is composed of a series of neural surfaces, each performing a different aspect of looking, recognising or pointing towards things. The important point is that all of these surfaces have a lot in common, despite the differences in their function: each can be described at some level as a servomotor, which performs some sort of coordinate transform and computes its results using the properties of large-scale patterns of nerve activity.

There’s an awful long way to go until these ideas gel into a unified and powerful mechanism capable of being used in real applications. But all adults start out as babies and I don’t see any reason why AI should be different. I’m quite convinced that there is such a universal architecture for creating mammal-like general intelligence, and that this bears little resemblance to existing neural networks and none whatsoever to the concepts underlying the digital computer. Until we find such a radical new way forward, I don’t think we will ever build robots for which Asimov’s three laws of robotics have the slightest relevance.

In closing, I’d like to point out that I’m sure many of the ideas I’ve outlined above have already been proposed in some form by other researchers. If so then I apologise for not citing them. My aim is not to plagiarise, it’s simply that I pay no attention to the literature. One of the best ways to kill off a promising line of thought is to say ‘well, I know that so-and-so tried that in 1978 and it didn’t work.’ In reality, the likelihood is that so-and-so didn’t have exactly the same ideas in mind, wasn’t thinking about them in precisely the same way and wasn’t driven by quite the same motives, so it’s better not to know. I’ve no doubt Edward de Bono would agree that most breakthroughs arise when people doggedly plough their own furrow, unknowingly attempting things that wiser people ‘know’ to be impossible. Ignorance can therefore be a huge asset, and as an unfunded amateur I have no obligation to stick to the rules and etiquette of professional science, so the less I know about other people’s ideas the better. In fact, if I were you I wouldn’t be reading this article at all – it would only colour my thoughts and reduce the potential for novelty. But perhaps the final paragraph wasn’t the ideal place to mention this.

 
Copyright © 2004 Cyberlife Research Ltd.
Last modified: 06/04/04