How Neural Networks Revolutionized AI
BROOKE GLADSTONE This is On the Media, I'm Brooke Gladstone. If you show a three year old a picture and ask them what's in it, you'll get pretty good answers.
3 YEAR OLD Okay. That's a cat sitting in a bed... Those are people going on an airplane. That's a big airplane.
BROOKE GLADSTONE Those are clips from a 2015 TEDTalk by Fei-Fei Li, a computer science professor at Stanford University. She was consumed by the fact that despite all our technological advances, our fanciest gizmos can't make sense of what they see.
FEI-FEI LI Our most advanced machines and computers still struggle at this task.
BROOKE GLADSTONE In 2010, she started a major computer vision competition called the Image Net Challenge, where software programs compete to correctly classify and detect objects and scenes. Contestants submit A.I. models that have been trained on millions of images organized into thousands of categories. Then the model is given images it's never seen before and asked to classify them. In 2012, a pair of doctoral students named Alex Krizhevsky and Ilya Sutskever entered the competition with a neural network architecture called Alex Net. The results were astounding.
GEOFFREY HINTON They did much better than the existing technology, and that made a huge impact.
BROOKE GLADSTONE Geoffrey Hinton was there, Ph.D. advisor at the University of Toronto and a collaborator on Alex Net. He had been working on a type of A.I. technology called neural networks since the 70s.
GEOFFREY HINTON It wasn't till the neural networks did really well, a vision that people really set up, and that kind of opened the floodgates.
BROOKE GLADSTONE His fascination with neural networks started when in high school, and a friend told him about holograms and the brain.
GEOFFREY HINTON Holograms had just come out. And he was interested in the idea that memories are distributed over the whole brain. So your memory of a particular event involves neurons in all sorts of different parts of the brain. And that got me interested in how memory works.
BROOKE GLADSTONE Hologram, meaning a picture or more, for lack of a better word, holistic way of storing information as opposed to just words. Is that what you mean?
GEOFFREY HINTON No, actually, a hologram is a holistic way of storing an image as opposed to storing it pixel by pixel.
BROOKE GLADSTONE Ah.
GEOFFREY HINTON So when you sort of pixel by pixel, each little bit of the image is stored in one pixel when you saw it in a hologram. Every little bit of the hologram stores the whole image. So you can take a hologram and cut it in half and you still get the whole image. It's just a bit fuzzier. It just seemed like a much more interesting idea than something like a filing cabinet, which was the normal analogy where the memory of each event is stored as a separate file in the filing cabinet.
BROOKE GLADSTONE There was somebody named Carl Lashley, you said, who took out bits of rats brains and found that the rats still remembered things.
GEOFFREY HINTON Yes. Basically, what he showed was the memory for how to do something isn't stored in any particular part of the brain. It's stored in many different parts of the brain. And in fact, the idea that, for example, an individual brain cell might store a memory doesn't make a lot of sense because your brain cells keep dying. And each time a brain cells as you don't lose one memory.
BROOKE GLADSTONE This notion of memory, this holographic idea was very much in opposition to conventional symbolic A.I.. Yes. Which was all the rage in the last century.
GEOFFREY HINTON Yes. You couldn't sort of draw a contrast between two completely different models of intelligence. In the symbolic AI model. The idea is you store a bunch of facts as symbolic expressions, bit like English, but cleaned up so it's not ambiguous. And you also store a bunch of rules that allow you to operate on those facts. And then you can infer things by applying the rules to the known facts to get new known facts. So it's based on logic, how reasoning works. And then they take reasoning to be the core of intelligence.
BROOKE GLADSTONE Hmm.
GEOFFREY HINTON It's a completely different way of doing business, which is much more biological, which is to say we don't store symbolic expressions. We have great big patterns of activity in the brain. And these great big patterns of activity which I call vectors. These vectors interact with one another. And these are much more like holograms.
BROOKE GLADSTONE So you've got these vectors of neural activity.
GEOFFREY HINTON So, for example, large language models that lead to big chat bots are all the rage nowadays. If you ask how do they represent words or word fragments? What they do is they convert a symbol that says it's this particular word into a big vector of activity that captures lots of information about the word that convert the word cat into a big vector, which is sometimes called an embedding. There's a much better representation of a cat than just a symbol. All the similarities of things are conveyed by these embedding vectors very different from a symbol system. The only property a symbol has is that you can tell whether two symbols are the same or different.
BROOKE GLADSTONE I'm thinking of more of the X paradox, which I understand is the observation by A.I. and robotics researchers that reasoning actually requires very little computation, but a lot of sensory motor and perception skills. He wrote in 88, It's comparatively easy to make computers exhibit adult level performance on intelligence tests or playing checkers and difficult or impossible to give them the skills of a one year old when it comes to perception and mobility, I just wonder, do you think machines can ever think until they can get sensory motor information built into those systems?
GEOFFREY HINTON There's two sides to this question: a philosophical side and a practical side. So philosophically, I think, yes, machines could think without any sensory motor experience, but in practice it's much easier to build an intelligent system if it has sensory input. There's all sorts of things you learn from sensory input, but the big language models that lead to these chat bots, many of them just have languages that are input. One thing you said at the beginning of this question was that reasoning is easy and perceptions hard. I'm paraphrasing. That was true when you used symbolic AI, when you tried to do everything by having explicit facts and rules to manipulate them. Perception turned out to be much harder than people thought it would be. As soon as you have big neural networks that learn and learn these big vectors, it turns out one kind of reasoning is particularly easy, and it's the kind that people do all the time, and it's most natural for people. And that's analogical reasoning.
BROOKE GLADSTONE Analogical reasoning. One thing is like another.
GEOFFREY HINTON Yeah, we're very good at making analogies.
BROOKE GLADSTONE So you went on to study psychology and your career in tech in which you are responsible for something that amounts to a revolution on A.I. Was an accidental spinoff of psychology. You went on to get a Ph.D. in A.I. In the seventies at the oldest A.I. Research Center in the U.K. That was the University of Edinburgh. You were in a place where everyone thought that what you were doing studying memory, has multiple stable states in a system wouldn't work. That in fact, what you were doing studying neural networks, was resolutely anti- A.I.. You weren't a popular guy. I guess.
GEOFFREY HINTON That's right. Back then, neural nets and I were seen as opposing camps. It wasn't until neural nets became much more successful than symbolic AI that all the symbolic AI people started using the term A.I. to refer to neural nets so they could get funding.
BROOKE GLADSTONE So when explaining the difference for a non-technical person between what a neural network is and why it was revolutionary compared to symbolic A.I., a lot of it hinges around what you think a thought is.
GEOFFREY HINTON I recently listened to a podcast where Chomsky repeated his standard view that thoughts and language are very close. Mm hmm. Whatever thought is, it's quite similar to language. I think that's complete nonsense. I think Chomsky is misunderstood how we use words. If we were to use computers and we had the same model of the world, then it would be very useful. One computer telling the other computer which neurons were active, and that would convey from one computer to another what the first computer was thinking. All we can do is produce sound waves or written words or gestures. That's the main way we convey what we're thinking to other people.
BROOKE GLADSTONE Mm hmm.
GEOFFREY HINTON A string of words isn't what we're thinking. A string of words is a way of conveying what we're thinking. It's the best way we have because we can't directly show them our brain states.
BROOKE GLADSTONE I once had a teacher who said. If you can't put it into words, then you don't really understand it.
GEOFFREY HINTON I think there were all sorts of things you can't put into words that your teacher didn't understand.
BROOKE GLADSTONE So the only place words exist is in sound waves and on pages.
GEOFFREY HINTON The words are not what you operate on in your head to do thinking. It's this big vector of activity. The words are just kind of pointers to these big vector of activity. They're the way in which we share knowledge. It's not actually a very efficient way to share knowledge, but it's the best we've got.
BROOKE GLADSTONE So today you're considered a kind of godfather of A.I.. There's a joke that everyone in the field has no more than six degrees of separation from you. You went on to become a professor at the computer science department at the University of Toronto, which helped turn Toronto into a tech hub. Your former students and post-doctoral fellows include people who are today now leading the field. What's it like being called the godfather of a field that rejected you for the majority of your career?
GEOFFREY HINTON It's pleasing.
BROOKE GLADSTONE And now all the big companies are using neural nets.
GEOFFREY HINTON Yes.
BROOKE GLADSTONE How do you define thinking? And do you think machines can do it? Is there a point in comparing A.I. to human intelligence?
GEOFFREY HINTON Well, a long time ago, Alan Turing, I think he got fed up with people telling him machines couldn't possibly think because they weren't human and defined what's called the Turing Test. Back then, you had teletypes and you would type to the computer the question and it would answer the question. This was a just sort of thought experiment. And if you couldn't tell the difference between whether a person was answering the question, whether the computer was answering the question, then Alan Turing said, You better believe the computer is intelligent.
BROOKE GLADSTONE I admire Alan Turing, but I never bought that. I don't think it proves anything. Do you buy the Turing Test?
GEOFFREY HINTON Um basically yes, it has problems with it, but it's basically correct. I mean, the problem is suppose someone is just adamantly determined to say machines can't be intelligent. How do you argue with them? Because nothing you present to them satisfies them that machines are intelligent.
BROOKE GLADSTONE I don't agree with that either. I could be convinced if machines had the kind of hologram like web of experience to draw from the physical as well as the mental and computational.
GEOFFREY HINTON The neural nets are very holistic. Let me give you an example from Chat GPT. There's probably better examples from some of the big Google models, but Chat GPT is better publicized. So you ask Chat GPT to describe losing one sock in the dryer in the style of the Declaration of Independence. It ends up by saying that all socks are endowed with certain rights, certain inalienable rights by their manufacturer. Now why did you say manufacturer? Well, it understood enough to know that socks are not created by God. They're created by manufacturers. And so if you're saying something about socks, but in the style of the Declaration of Independence, the equivalent of God is the manufacturer and understood all that because it has sensible vectors, the represents socks and manufacturers and God and creation. That's an example of a kind of holistic understanding, an understanding via analogies that's much more human like than symbolic A.I, and that is being exhibited by Chat GPT.
BROOKE GLADSTONE And that, in your view, is tantamount to thinking. It is thinking.
GEOFFREY HINTON That's intuitive thinking. What neural nets are good at is intuitive thinking. The big chatbots aren't so good at explicit reasoning, but they mirror people. People are pretty bad at explicit reasoning.
BROOKE GLADSTONE We don't have identical brains. Our brains run at low power about 30 watts right? And they're analog. We're not as good at sharing information as computers are.
GEOFFREY HINTON You can run 10,000 copies of a neural net on 10,000 different computers, and they can all share their connection strings because they all work exactly the same way. And they can share what they learn by sharing their weights, their connection, strengths to computers that are sharing a trillion weights. It's an immense bandwidth of information between the two computers, whereas two people who are just using language have a very limited bottleneck.
BROOKE GLADSTONE So computers are telepathic.
GEOFFREY HINTON It's as if computers are telepathic, right.
BROOKE GLADSTONE Were you excited when Chap GPT was released? We've been told it isn't really a huge advancement. It's just out there for the public.
GEOFFREY HINTON In terms of its abilities. It's not significantly different from a number of other things already developed, but it made a big impact because they did a very good job of engineering it, so it was easy to use.
BROOKE GLADSTONE Are there potential implementations of AI that concern you?
GEOFFREY HINTON People using A.I. for autonomous lethal weapons. The problem is that a lot of the funding for developing A.I. Is by governments who would like to replace soldiers with autonomous lethal weapons. So the funding is explicitly for hurting people. That concerns me a lot.
BROOKE GLADSTONE That's a pretty clear one. Is there something subtler about potential applications that give you pause?
GEOFFREY HINTON I'm hesitant to make predictions, beyond about five years, it's obvious that this technology is going to lead to lots of wonderful new things. As one example, Alpha Fold, which predicts the 3D shape of protein molecules from the sequence of faces that define the molecule, that's extremely useful and is going to have a huge effect in medicine. And there's going to be a lot of applications like that. They're going to get much better at predicting the weather, not beyond like 20 days or so, but predicting the weather in like ten days time. I think these big AI systems are already getting good at that. But there's just going to be huge numbers of applications. In a sensible society, this would all be good. It's not clear that everything's going to be good in the society we have.
BROOKE GLADSTONE What about the singularity? The idea that what it means to be human could be transformed by a breakthrough in artificial intelligence or a merging of human and artificial intelligence into a kind of transcendent form?
GEOFFREY HINTON I think it's quite likely we'll get some kind of symbiosis. A.I. Will make us far more competent. I also think that the stuff that's really happened with neural nets is changing our view of what we are. It's changing people's view from the idea that the essence of a person is a deliberate reasoning machine that can explain why it arrives at conclusions. The essence is much more a huge analogy machine that's forever making analogies between a gazillion different things to arrive at intuitive conclusions very rapidly. And that seems far more like our real nature than reasoning machines.
BROOKE GLADSTONE Have you ever had a flight of fancy of what this ultimately might mean and how we live?
GEOFFREY HINTON That's beyond five years.
BROOKE GLADSTONE You're right, I see. You warned me! Jeffrey, thank you very much.
GEOFFREY HINTON Okay.
BROOKE GLADSTONE Geoffrey Hinton is an engineering fellow at Google Brain.
Coming up with great computing power comes great responsibility. This is On the Media.
Copyright © 2023 New York Public Radio. All rights reserved. Visit our website terms of use at www.wnyc.org for further information.
New York Public Radio transcripts are created on a rush deadline, often by contractors. This text may not be in its final form and may be updated or revised in the future. Accuracy and availability may vary. The authoritative record of New York Public Radio’s programming is the audio record.