[Chatbot]
Write a short introduction for Geoffrey Hinton, the subject of this video. Geoffrey Hinton is a University of Toronto professor emeritus who is
known as the "Godfather of AI." He recently left Google so he could
more freely discuss the dangers posed by unchecked AI development. We spoke to him in his London home
about the technology he helped create, its many benefits and why he
suddenly fears humanity is at risk. [Geoffrey Hinton] I got a request from the Wall Street Journal. They want me to correct my obituary. [Christopher Sorensen]
What do you mean? [Hinton]
They want me to correct my obituary! [Sorensen]
They've like pre-written it, right? [Hinton]
They've pre-written my obituary. <laughter> I wonder what Mark Train
would've said about that. [Sorensen]
So. I guess we don't really need an introduction here
so I will just launch right into it. You've recently given a
number of interviews in which you've said
that digital intelligence that is used by chatbots
and other generative AI may be a better kind of intelligence
than the biological intelligence that we have. Can you briefly explain what
made you come to this conclusion? [Hinton]
So, in a digital computer it's designed so
you can tell it exactly what to do and it'll do exactly what you tell it. And even when it's learning stuff, two different digital computers can do
exactly the same thing with the same learned knowledge. And that means that you could
make 10,000 copies of the same knowledge, have them all running
on different computers and whenever one copy
learns something, it can communicate it very
efficiently to all the other copies. So you can have 10,000
digital agents out there, a kind of hive mind, and they can share knowledge extremely
efficiently by just sharing the connection strengths inside the
neural nets. And we can't do that. If you learn something and
you want to tell me about it, you have to use sentences or pictures, and you can only share a very limited
amount of information that way. So it's much, much slower for you to communicate
what you've learned to me than it is for these digital intelligences
to communicate stuff. And that makes them much better. They can learn a whole
lot of stuff between them. [Sorensen]
You've said that digital intelligence is immortal and that biological
intelligence is mortal. What did you mean by this? [Hinton]
So if I learn some connection strengths in a neural net that's being
simulated on digital computers, then if a particular computer dies, those same connection strengths
can be used on another computer. And even if all the
digital computers died, if you'd stored the
connection strength somewhere, you could then just make another digital
computer and run the same weights on that other digital computer. But with us, the knowledge that we learn,
the connection strengths, are specific to our particular brains.
Every brain is a bit different. The neurons in your brain are all a
bit different and you learn so as to make use of all the idiosyncrasies
of your particular brain. And so once you've learned
connection strengths in your brain, if you told me those
connection strengths, they wouldn't do me any good
because my brain's different. So the digital computers are immortal
because you can run that same knowledge on a different piece of hardware. We are mortal because the hardware and
the knowledge are intricately entangled. You can't separate the connection strengths from
the particular brain they're running in. And so if the brain dies,
the knowledge dies. [Chatbot]
Why should we be concerned about digital intelligence taking over
from biological intelligence? [Hinton]
Because I think it's much better at sharing what's learned by a whole
bunch of different digital agents who all share the same weights and
they just share the updates to the weights, and now they can learn 10,000
different things at the same time. But also I think the digital intelligence
probably has a better learning algorithm than the brain's got. All the attempts to find a learning algorithm
in the brain that works as well as the
backpropagation algorithm in these digital intelligences —
so far those attempts have failed. We haven't found anything that scales
up as well to very large systems as the backpropagation algorithm. So I think they've got
two advantages. They've probably got a better learning
algorithm and they can share knowledge much more efficiently than
biological intelligences can. [Sorensen]
At the time when you entered the field
there were two schools of thought on
machine intelligence: mainstream and neural nets. Can you describe the difference
between these two approaches? [Hinton]
I can sort of caricature it. So there's two different models of
what intelligence is all about and one model is that it's all about reasoning
and the way we reason is by using logic. And so that's what's
special about people. And what we should be doing is understanding the kind of logic
that we actually use. And that also went with
the idea that the knowledge you store is
symbolic expressions — so that I can say a sentence to you and you will somehow
store that and then later you'll be able to use it for
inferring other sentences. But what's inside your head is something
a bit like sentences but cleaned up. And there's a completely different model
of intelligence which is that it's all about learning the connection
strengths in a network of brain cells. And what it's good for is things
like perception and motor control, not for reasoning.
That reasoning came much, much later and we are
not very good at it. You don't learn to do it
until you're quite old. And so reasoning is actually a very
bad model of biological intelligence. Biological intelligence is
about things like controlling your body
and seeing things, and that was a totally different paradigm
and had a different idea of what's in inside your head — that it's not stored strings of symbols,
it's just connection strengths. In the symbolic AI view, the crucial question was: What is the form of these
symbolic expressions and how do you do the
reasoning with them? For the neural net view, the central question was
quite different. It was how do you learn these connection
strengths so you can do all these wonderful things? And so learning was
always central to the neural net view. For the symbolic view, they said,
"We'll worry about learning later." "First you have to figure out how the
knowledge is represented" "and how we reason with it." And so these were
totally different views. One took its inspiration from
logic and one from biology. And for a long time the people in the
logic camp thought taking your inspiration from biology was silly. That was a bit strange
since von Neumann and Turing had both thought neural nets were
the way to attack intelligence. But unfortunately they
both died young. [Chatbot]
Can you at a high level describe how a neural network works? [Hinton]
I can try. So let's start off by describing
how it would work for recognizing objects and images. And let's suppose all we wanted to do was say whether or not
there was a bird in the image. And let's suppose a bird's gonna be
roughly in the middle of the image and the main object of attention and you have
to say, "Is this a bird or isn't it?" So you can think of an image. Let's suppose it's a hundred pixels by
a hundred pixels. That's 10,000 pixels. Each pixel is three colors:
RGB — so that's 30,000 numbers. And in computational terms, recognizing a bird in an image
consists of taking 30,000 numbers and outputting one number that
says yes or no, it's a bird. And you could try and write a
standard computer program to do that. And people tried for many, many years, and they could never get it
to work very well. Like for 50 years they were trying to do that. Or you could make a multilayer
neural net. And I'll start off by telling you how
you would wire up a neural net by hand. So what you'd do is you'd
have the pixels and that will be the bottom level. And then you'd have a
layer of feature detectors, and a typical feature detector
might have big positive connection strengths coming from a vertical row
of pixels and big negative connection strengths coming from a
neighbouring vertical row of pixels, and no connection strengths
anywhere else. So if both rows of pixels are bright, it'll get big positive input from here,
but also big negative input from there. So it won't do anything. But if these ones are bright,
giving a big positive input, and these ones are not bright,
so it doesn't get inhibited by these ones, it will get all excited and it will
say, "Hey, I found the thing I like," "which is bright pixels here
and dark pixels here." And that's an edge detector. I just told you how to wire up by hand
using positive and negative weights, something that would detect
a little vertical edge. So now imagine you have a gazillion of
those guys detecting different edges in different locations in the image, in different orientations
and at different scales. That will be your first-layer
feature detectors. Now if I was wiring it by hand my
second-layer of feature detectors, I would maybe have a detector
that takes two edges of joint at a fine angle like this. So it's
looking for this edge and this edge. And if they're both active
at once it would say, "Hey, maybe there's a beak here." It could be all sorts of other things
but it might just be a beak. So you have a
feature that's sort of beak-like. You might also have a feature that detects a whole
bunch of edges in the form of a circle. And so you'd have circle detectors
and potential beak detectors, as well as lots of other
detectors in that layer. But they're detecting slightly
more complicated things. And then in the layer above that you
might have something that detects a potential beak in the right spatial
relationship to a potential circle, a potential eye, so that it
could be the head of a bird. So that would be like your third layer. And maybe if in your third layer you
also got something that detected the foot of a bird and the wing of a bird, then maybe in the next layer you could
have a bird detector - that if several of
those things got active, like OK, here's a head and there's a wing and
there's a foot, it probably is a bird. Ok, so I told you how to wire all those things
up by hand but you'd never be able to do a very good job of it. So, instead of wiring it all up by hand, we
could imagine trying to learn it all. So I've told you the kind
of thing we want to learn, but now I'll tell you how we learn it and the way we learn it
sounds bizarre at first. Instead of wiring in all the connection
strength so you get the detectors you want, you start with random connection
strengths — just random numbers
on all the connections. And so you put in an image of a bird and
you go forward through these layers of feature detectors and it just
behaves completely randomly. And the bird detector at the
output will say, 0.5 it's a bird. It's gonna say one when it's sure it's
a bird and zero and it's sure it's not a bird, so it's going to say about 0.5. And now you can ask
the following question: how can I change all those
connection strengths in the network, so instead of saying 0.5 it's a bird —
let's suppose it is a bird - it says 0.51 it's a bird. So the question you want to ask is: how should I change a particular
connection strength so as to make it more likely that it's a bird? And you can figure that out by
taking the difference between what you got and what you wanted. So you
wanted one and you actually got 0.5. You take that difference and you send that difference
backwards through the network. And then you use some
calculus, which I won't explain. And you are able to compute for every
single connection in the network how much you'd like to make it
bigger or smaller in order to make it more likely to say "bird." Then you adjust all the connection
strengths very slightly in the direction that are making
it more likely to say "bird." Then you show it something
that isn't a bird and now you're gonna adjust
connection strengths so it's less likely
to say that that was a bird. And you just keep going like that
with lots of birds and non-birds, and eventually you'll discover that
it's discovered all these feature detectors - it'll have discovered
beak-like things and eye-like things, and things that detect feet
and wings and all that stuff. And if you train it on
lots of different objects, like a thousand different
categories of object, it'll discover intermediate
feature detectors that are very good for
recognizing all sorts of things. So the magic is that there's this
relatively simple algorithm called backpropagation that takes the error
in the output and sends that error backwards through the network and computes
through all the connections how you should change them to
improve the behavior, and then you change them all a tiny bit and you just keep going
with another example. And surprisingly
that actually works. For many years people thought that would
just get jammed up — it would get stuck somewhere — but no it doesn't,
it actually works very well. [Chatbot]
I'm curious, how do neural
networks handle language? [Hinton]
OK, so now you've got the idea of how we train it to recognize a bird. Imagine now that we take a string of words as the input. The first thing you are going do is
convert a word into an embedding vector — that is, it's a little bunch of numbers
that captures the meaning of the word, or is intended to capture
the meaning of the word. And so your first layer
after the words will be these embedding vectors
for each word. And now we're gonna have lots of layers
of embedding vectors and as we go up through the network, we're gonna make the embedding vectors
for a word get better and better because they're going to take into account
more and more contextual information. So suppose in this sentence — let's
suppose we don't have any capital letters, OK? So suppose in this sentence
you have the word "may." Well, the most probable meaning of "may" is
that it's a modal as in, "He may do that." But obviously there's a completely
different meaning of "may," which is the month. And so initially it doesn't know, just looking at the word "may," it
doesn't know what embedding vector to use. And it'll use a kind of compromise vector, something that's sort of halfway between
the embedding vector that represents the modal "may" and the embedding
vector that represents the month "May." And then at the next layer
it's going refine that vector. It's going to make a
slightly better vector depending on the context that it got,
depending on nearby embedding vectors. So if, for example, nearby there's the
embedding vector for "June," then it'll refine the one for "may" to
be more like a month and less like a modal. But if there's the embedding vector for
"would," it'll make it more like a modal and less like a month. And as you go
through the network, it can refine these embedding vectors
and make them better and better. And the way we are going to train it, is we're going to give it
a string of words as input and we are going to — here
will be one way to do it ... it's not exactly what's done but it's
easy to understand — for the last word, you just put in a kind of neutral word. You say "unknown" and it has a very
vague embedding vector that's kind of the average of all the vectors for all
words. It doesn't know, right? Now, as you go forward
through the network, that last word will be able to
be influenced by previous words, and it starts off very vague, but as you go through these layers
it can get more and more precise. And by the time you get to
the end of the network that embedding vector could look like the
embedding vector for a particular word or for some combination
of words — some average of several words. And you train
the network by saying, "You go through all these layers and
that last word you'd like the embedding vector to look like the embedding vector
for the word that actually was there in the text." And that's how
it predicts the next word. It tries to change this sort of neutral
embedding vector into one that is close to the embedding vector for the
correct word that appeared in the text. And you take the error, the difference between the
embedding vector in the text and the embedding vector produced, and you propagate that backwards through
the network — and it's propagating
backwards through the layers. but it's propagating from this word to
previous words so that they will have the right influence on this word. And that's the backpropagation algorithm
learning to predict the next word. [Sorensen]
So despite some of the theoretical
breakthroughs in this field, these neural networks didn't work very
well for a long time. And why was that? [Hinton]
It was a combination of reasons. So we weren't very good at
initializing them — that, as I said, you put in random weights
and then learn everything. But if you don't carefully decide
what kind of random weights, the thing never gets off the ground. So that was a little technical reason
why they didn't work very well in deep nets with lots of layers
of feature detectors. But the main reason was we didn't have
enough compute power and we didn't have enough data. So people were trying to train these
nets on relatively small training sets without much compute power. And in
that regime other methods work better. Neural nets really come into their own
when you have a lot of data and a lot of compute power, and then you can use a big neural net
and then it works much better than anything else. And we didn't realize that at the time.
So we would occasionally fantasize, "Well, suppose you had a lot more data and
a lot bigger in computer it will work better." But we didn't realize
it will work a whole lot better. And so in the 1990s, it was a relatively dead period for
neural nets because other methods were working better on small problems. And a lot of people in computer
science gave up on neural nets. In psychology they didn't,
because in psychology they wanted something that was
like the brain and neural nets were clearly more like the brain
than symbolic AI. But in computer science,
neural nets sort of came
into disrepute in the 90s. [Sorensen]
So let's fast forward then to another decade, to the 2000s. Was there a moment for you when it
became clear that the approach that you'd been pursuing was the one
that was going to prevail? [Hinton]
OK. In 2006, we figured out how to initialize the
weights much better by doing unsupervised learning and then backpropagation
worked much better. So it was fairly clear then the backpropagation really
was going to work very well. But in 2009, two of my grad students — George
Dahl and Abdelrahman Mohamed — made a much better speech recognizer, actually a slightly
better speech recognizer, but it was slightly better than the
state of the art, using deep neural nets. And then it was fairly clear that
this stuff was going somewhere. And all the big speech groups over the
next few years switched to using neural nets. And then in 2012, that speech stuff
came out in the Android and suddenly the Android caught up with Siri. It was as good at speech as
Siri because it was using neural nets. And in the same year two others
of my graduate students — Ilya Sutskever and Alex Krizhevsky — made a neural net that
was very good at recognizing objects and images and that beat the
state of the art by a lot. And so I think it was this combination
that it was already working for speech recognition and already in production
— the big companies knew that, but the public I don't think were very
well aware of that — but then suddenly it worked much
better for computer vision and that was a turning point. In 2012, when we won the ImageNet
competition by a huge margin, we got almost half the errors of
the other methods and it was a public data set but with a hidden
test set, so you couldn't cheat. [Sorensen]
So let's just focus a bit on 2012 because you said it was a
really pivotal year for this. Can you describe again at
a high level how AlexNet worked? I take it that might have been
named after your graduate student. [Hinton]
That was named after Alex Krizhevsky because he did ... he was a
wizard programmer and he made it work. Ilya helped a lot but it
was mainly Alex's work. So I explained to you when explaining
backprop how you'd have these layers of feature detectors and AlexNet
was basically that kind of a net but with a thousand different
object classes and with about seven layers of feature detectors. And it also used something else
that was developed by Yann LeCun, which is convolutional nets. And I'll try and explain those now
because they were very important. Remember how I said you might make
a detector for a bird's beak by checking two lines ...
by having two lines like that? And if you see those two feature
detectors then you make a beak detector? But that would just be for a specific
location right? In a convolutional net, when you make a feature
detector for one location, you make the same feature detector
for all the locations in the image. So now if it's trained with a
beak here when it's learning and it really says, "I need a
beak detector for that" — so, it learns a feature that detects this
beak — it will automatically make copies for all of the other
locations in the image. So if now the bird occurs
in a different location, it will have the feature
detectors to recognize it. So that idea that you copy the
feature detectors to every location, that's a convolutional net essentially. And that makes the whole thing
generalized much better across position. It can cope now with things changing
position because it's got copies of all these feature detectors in every location. And with convolutional nets
and multiple layers of features, what Alex did was programmed all that
very efficiently on a thing called a graphics processing unit, which was
developed for computer graphics. But it's like a mini super computer, it can do lots and lots of computation
in lots of separate processs all at the same time. And so it gave us about a
factor of 30 compared with a normal computer and a factor of 30 is about sort
of 10 years progressing in computers. So suddenly we could leap 10 years
into the future in terms of compute power. And it was very difficult to program
these GPU boards. Alex managed to
program two of them to collaborate, which was even more difficult. And the last ingredient
was the ImageNet dataset. So someone called Fei-Fei Li and her
collaborators put together a big set of images and then a public competition
where you had about a million images with a thousand different
kinds of objects. So you had about a thousand examples of
each kind of object and you had to learn to recognize those objects. And then the test set
would be different images, which also contain those objects. And so you'd have to generalize
to the different images. And it turned out the best computer
vision technique that had been invented up until then was getting
like 25 per cent errors and Alex got 15 per cent errors. And since then it's gone down
to about 3 per cent errors. It's got much better since then. But it was a huge jump and people in
computer vision were extremely surprised and most of them behaved in a very
admirable way, which is they said, "Hey, we never thought this would work," but hey it works so we're gonna do that
instead of what we were doing." That's what scientists don't usually do. Scientists usually just grow old complaining that
this new stuff is nonsense. [Sorensen]
And how would you describe the pace of innovation that we've
seen in AI since that moment? [Hinton]
It's just got faster and faster. So if you'd asked me in that moment
how long until these neural nets can do machine translation that's
better than the state of the art, I'd have said maybe 10 years. Because machine translation
is the kind of thing that, if you've got a theory that's all
about processing strings of symbols, machine translation is the ideal problem
for you because you have a string of symbols in one language and you have to produce a string
of symbols in another language. And the symbolic people thought, well inside you're just
manipulating strings to do that. The neural net people thought you
have to take this string of symbols, you have to convert it into these big
patterns of neural activity and then you have to convert it back
into symbols at the output. And I was very surprised when it only
took a few years for machine translation to be good and then in another year or
two Google was using it and it greatly improved the quality
of machine translation. Like in languages like Chinese —
this is from memory, but ... — there was a gap between how good the
computer translation was and how good human translation was.
And it just halved that gap overnight. I think it was Chinese that it did that, but in a lot of languages it just
made it a lot better. And since then, it's obviously it's got
considerably better since then. But by 2015 it was already working pretty
well and that really surprised me that it only took three years. [Chatbot]
You say you were surprised at the pace of innovation. What did you think the first time you
used a large language model like ChatGPT? Did we surprise you? [Hinton]
I am just shocked at how ...
how good it is. So it gives very coherent
answers and it can do little bits of reasoning — not very
sophisticated reasoning yet, although it'll get much
better. So, for example, I asked it — this is
GPT-4 now — I asked it a puzzle given to me by a symbolic AI guy
who thought it wouldn't be able to do it. I actually made the puzzle much
harder and it could still do it. And so the puzzle goes like this: the rooms in my house are
either white or blue or yellow. Yellow paint fades to white within a year. In two years time, I would like all the
rooms to be white, what should I do? And a human being will probably say
you should paint the blue rooms white. What GPT-4 said was you should paint the
blue rooms yellow — but that works too because yellow will fade to white. And I don't see how it could do that
without understanding the problem. The idea that it's just sort of
predicting the next word and using statistics — there's a sense in which that's true, but it's not the sense of statistics
that most people understand. It, from the data, it figures out how to extract the
meaning of the sentence and it uses the meaning of the sentence
to predict the next word. It really does understand
and that's quite shocking. [Sorensen]
So have you been surprised by the broader reaction,
the public reaction, to ChatGPT? Well given how well it works, I guess the
public reaction isn't that surprising. But what's interesting is,
most people don't say, "This doesn't understand." They say, "Wow, it understood what I
said and gave me a coherent answer." "What can I use it for?" And I think most people are right about
that and of course it can be used for huge numbers of things. So I know someone who answers letters
of complaint for the health service, and he used to spend 25 minutes composing
a letter that addresses the problem. And so now he just types the problem into GPT-4 and it writes the letter and
then he just looks at the letter and decides if it's OK and sends it out.
And that takes him five minutes now. So he is now five times more efficient and that's gonna
happen all over the place. Like paralegals are
gonna be like that. Programmers are already
getting like that. Programmers can be much more efficient
if they get assistance from things like GPT-4 because it
knows how to program. And you might think it just knows how
to program because he's seen a whole lot of programs. So I have a former graduate student who's
very smart and a very good programmer and he did a little experiment which
is — he's called Radford Neal — he took GPT-4 and he defined a new programming
language with very unusual syntax, and having defined this programming
language just in text to GPT-4, he then gave it a program and said, "What would this do?"
And it answered correctly. So basically it could understand the
definition of a new programming language and figure out what programs in
that language would do. And again, the idea that it's just predicting
the next word doesn't make any sense
in that context. It had to understand
what was going on. [Sorensen]
So what do you see as some of the most promising opportunities for this
type of AI when it comes to benefiting society? [Hinton]
It's hard to pick one because there's so many — Like there'll be a huge increase in productivity for any
job that involves outputting text. There's all sorts of issues about
increasing productivity. In our society, it's not necessarily a good
thing to increase productivity because it might make the
rich richer and the poor poorer. But in a decent society, just increasing productivity
ought to be a good thing. So there'll be things like that. It's wonderful for making predictions. It'll be better at predicting the weather. People don't know by how much yet, but it's already much
better at predicting floods. It can predict earthquakes.
It can design new nanomaterials. So for things like solar panels, you want to be able to design new
nanomaterials, or for superconductivity. I don't know if it's used
for superconductivity yet, but it may well be — you'd
like that at high temperature. It's really good at
designing drugs — that is, finding molecules that'll bind to
some particular other molecule. DeepMind has used it to create AlphaFold. Now that's not a chatbot.
That's just deep learning. But the basic technology
of deep learning has pretty much solved the problem
of how you figure out from the string of bases in a protein,
what shape it will adopt. And if you know what shape it
adopts, you know its function. The chatbots are just gonna
be used everywhere I think. [Sorensen]
And we've also talked a lot about health care. I mean, you
talked about drug discovery, but health care is another field
that could really benefit. [Hinton]
Yes. Both in interpreting medical scans ... Like if you take a CAT scan, there's a lot of information in
the CAT scan that isn't being used and most doctors don't know
what the information is. This will be able to get much more out
of a CAT scan as well as being able to compete with doctors at saying what kind of cancer you have
or how big it's grown. At present for example, when a doctor tells you
the size of a cancer, you'll get a number like it's
three centimetres and a month ago it
was two centimetres. Now that's not a very useful number
if the thing looks like an octopus, right? A neural net will be
able to do much better at understanding the volume of the
cancer and how it's changed. So it's going be tremendous there and
already it's at the level of humans for lots of kinds of scans and it's going
to get better. It's going to be very good
for diagnosing diseases. So at present, there's a large number of people who
die in North America because the doctors misdiagnosed what they had. There's a system that Google's
producing called Med-PaLM 2 which has learned to do diagnoses
and it's already ... I think it's better than
an average doctor now. I'm not quite sure about this because
I'm not at Google anymore and it's very recent, but it's certainly comparable with
doctors and it's going to get better fast. So wouldn't you like to have a
sort of general practitioner family doctor, and you go with some rare
disease and you'd love your family doctor to have already seen hundreds of
cases of that rare disease and Med-PaLM 2 is going to be like that. So it's going to be just, in the end,
much better at diagnosis. [Chatbot]
It sounds like AI will bring many important benefits, but you have expressed concern about
the current pace of innovation. Why? [Hinton]
OK, so for like 50 years I thought that — well for 49 years —
in order to make digital models better, we needed to make them
work more like the brain. So I kept looking at things the brain
does and the digital models don't like rapidly changing connection
strengths in a temporary way, and that can make the
digital models better. And very recently I realized that because
these digital models have this kind of hive mind where when
one agent learns something, all the other agents know it, they might actually already be
better than biological intelligence. And so I kind of completely flipped my
opinion from the idea it's going to be a long time before they can do everything
the brain does — that it's gonna be 30 to 50 years
before they're better than us, which is what I thought for until
very recently. A few months ago, I suddenly realized maybe
they're already better than us. They're just smaller and
when they get bigger then they'll be smarter than us. And that was quite scary. It was a sudden change of opinion
that instead of being 30 to 50 years, it was five years to 20 years, or something like that — and so we needed now to
take really seriously right now what we're going to do about the issue
that these things may become smarter than us. It's a time of huge uncertainty. Nobody really
knows what's going to happen. Maybe things will stall and maybe
they won't become smarter than us, but I don't really believe that. I think
they're going to be smarter than us. But maybe when they
become smarter than us, we'll be able to keep them benevolent
and we will we'll be able to keep them caring much more about people than they
care about themselves, unlike people. But maybe not. And so we need to start thinking very
hard about those issues and I'm not an expert on those issues. I'm just an expert on these
learning algorithms and I suddenly realized these superintelligences
may be here quite soon — and I'm just sounding the alarm so that people
listen to the experts who've been thinking for a long time about how
we might stop them taking control. I want the politicians to listen to
those guys rather than say, "Yeah, yeah, they're sort of sci-fi guys, they're
... it's never gonna happen.". [Sorensen]
Was there like a particular moment when you had this —
you said it was very recent — where you kind of
changed your view on it? I was developing learning algorithms
for biological systems that could run in a biological system, which didn't use backpropagation. And I couldn't make them work as well
as a backpropagation algorithm that we were running in these digital systems. And they would work for small networks. But when I scaled it up, the digital ones always scaled up
much better than the biological ones. And suddenly I thought
it might not be my fault. It might not be that my learning algorithm
was just a bad learning algorithm. It might be that these digital
systems just are better. And that's when I suddenly changed
my mind about how long before we get superintelligence. And then I talked to various
former students of mine, and former colleagues of mine, and some of them encouraged me
to go public with this — not because I had any solutions
that I wanted to recommend. It's not like you can say, "Burn less carbon and
everything will be fine." But because they thought I'm well
known in the field and if I go public by saying superintelligence
might be here quite soon, the politicians might start to believe
that's a possibility and start listening seriously to the researchers who've
been thinking a long time about how we prevent these things from gaining control. [Sorensen]
So from your point of view, what can ... what role can governments
play in helping ensure AI are developed in a responsible way? [Hinton]
So there's all sorts of risks other people have talked about a lot — and that I don't particularly
want to talk about — like they'll take jobs away and increase the gap between
the rich and the poor; they will make it impossible to
know whether the news is fake or real; they will encourage society
to divide into two warring camps that don't listen to each other
and have completely opposing views; they will build battle robots
that are designed to kill people. All of those are well known
risks that I'm not talking about. It's not that I don't
think they're important, I think they're probably
even more urgent, but lots of other people are
talking about those risks. The risk I'm talking about is the risk
that these things will get smarter than us and eventually take over. And for that risk there may be
something governments can do because nobody wants that ... well, if you exclude these super
intelligences. No people want that. And so all the different
governments ought to be able to agree ... they ought to be able to work together
on preventing that because it's in their interests. And that's happened
before, even during the Cold War. The U.S. and Russia could work together
on trying to prevent there being a global nuclear war because
it was so bad for everybody, and for this existential threat, it should be possible for everybody
to work together to limit it if it's possible to prevent it. I don't know whether it's
possible to prevent it, but at least we should be able to get
international collaboration on that particular threat: the existential
threat of AI taking over. One thing I think should be done
is wherever this stuff's being developed, particularly
these big chatbots, governments should encourage the
companies to put a lot of resources, as these things are getting
more and more intelligent, to doing experiments to figure out how
to keep them under control. So they should be sort of looking at how these
things might try and escape and doing empirical work on that, and put a lot of resources into that
because that's the only chance we've got. Before they're ... before
they're super intelligent, we can maybe do experiments and
see what's going to go wrong. And I'm strongly of the belief
you need empirical data on this. You just can't have philosophers
and politicians and legislatures making up rules. You need empirical work looking at these
things and seeing how they go wrong and seeing how you might control them. And that can only be done by
the people developing them. So since you can't stop the development, the best you can do is somehow have
governments put a lot of pressure on these companies to put a lot of
resources into investigating empirically how to
keep them under control when they're not
quite as smart as us. [Sorensen]
And what do you see as the role of these big technology companies where a lot of this development
is happening? Would they do this without that
kind of government regulation? [Hinton]
So a lot of the people in the big companies — all the people I know who are senior in the
big companies are very worried about this and do put work into that. They're very concerned about it, but they have an obligation to
their shareholders, and I think, to make big profits — and making big profits, particularly
in the short term, doesn't align nicely with putting a lot
of effort into making sure it's safe. So you see this in all industries.
In the railway industry in the States, having safety devices that tell you
when a wheel's locked costs money and the big rail companies just
rather have accidents than do that. Google, which is a big company
I know something about, is not quite like that because it
understands that it's going to have a tremendous reputational
loss if bad things happen. And that's why Google didn't release
these chatbots. It kept them private. It didn't want them out there in
the world for people to play with. It wanted to use them to
give you better search results or to complete your Gmail for you, but not to give them
to people to play with. And it could only be responsible like
that until OpenAI and Microsoft put them out there and then Google had to compete. But the big people in the big companies
really care a lot about their reputation and about not having bad effects, but they could maybe be made to care
even more about this safety issue by government doing something to insist
that they put a lot of work into that. And there's other things
that could happen. Like it's very hard within a company
to have people working on long-term existential threats
because they're paid by the company and there's a conflict of interest, which is one of
the reasons I left Google — not because Google did anything wrong because I just don't
want any conflict of interest. One thing the big companies
could certainly do is put more money into funding
foundations that study these things and Google, for example,
put $300 million into a foundation called Anthropic that is studying these things. They could put a lot more money in. [Sorensen]
I'm curious about what advice you would give,
or what guidance you would give, to other researchers in the field — you know, who might be just
entering the field right now and want make sure that they're advancing the field but doing
it in a responsible way? [Hinton]
Well, one piece of advice I'd give is look at how many people are
working on making these things better and how many people are working
on preventing them from getting out of control. And you'll see it's like 99 people are
working on making them better and one person's working on preventing them
getting out of control. So where could you
make the most impact? Probably on working in preventing
them getting out of control. So that's one piece of advice. The other piece of advice is my
general advice for young researchers, which is look for
somewhere where you think everybody's doing it wrong and trust your intuition until you figure
out why your intuition is incorrect. Trust it and work on alternatives ... alternative ways of doing
things when you think everybody else is doing it wrong. And the fact is either you
have good intuitions or you don't. If you've got good intuitions, you should listen to them and follow
your intuition and work on that until you discover why it's wrong. If you've got bad intuitions, it doesn't really matter what you do. So you might as well
follow your intuitions. [Chatbot]
The risks you've described are alarming, but can't you just throw
a switch and shut it down? Aren't humans ultimately still in control? [Hinton]
It's very tempting to think we could just turn it off. Imagine these things are a lot smarter
than us and remember they'll have read everything Machiavelli ever wrote. They'll have read every example in
the literature of human deception. They'll be real experts at doing human
deception because they'll have learned that from us, and they'll
be much better than us. They'll be like you
manipulating a toddler. You know, you say to your toddler, "Do you want peas or cauliflower?" And
your toddler doesn't realize actually he doesn't have to have either. He just thinks which he likes and which he dislikes the most
and says he'll have the other one. So if they can manipulate people, they can manipulate people into
pressing buttons and pulling levers. So we have a nice
example of Donald Trump. Donald Trump can manipulate people
and so he could invade a building in Washington without ever
going there himself. And you didn't have to prevent Donald
Trump from doing anything physical. You'd had to prevent him from talking
to prevent that. And these are chatbots. So the idea that just with talk, they can't do any real damage because
it requires people to do the damage — well, as soon as you
can manipulate people, then you can get
whatever you like done. [Chatbot]
You've spent your career trying to understand how the human brain works and played a critical role in AI development.
What's next for you, Geoffrey Hinton? [Hinton]
OK, so I'm 75 and I've reached the point where I'm not
very good at writing programs anymore because I keep forgetting the
names of the variables I'm using and things like that. And I forget to, I do a copy and paste and forget
to modify the thing I pasted. And so I've slowed down a lot in
programming and it's very irritating. It's extremely irritating not
to be as good as you used to be. And I decided a long time ago
that when I reached that point I would become a philosopher. And so I'm going to
become a philosopher.