TRANSCRIPTEnglish

Is AI Hiding Its Full Power? With Geoffrey Hinton

1h 33m 26s17,529 words2,709 segmentsEnglish

FULL TRANSCRIPT

0:00

Are we at a point where the artificial

0:02

intelligence will play down how smart it

0:04

is?

0:05

>> Yes. Already we have to worry about

0:06

that. If it senses that it's being

0:09

tested, it can act dumb.

0:10

>> What did you just say?

0:11

>> The AI starts wondering whether it's

0:14

being tested. And if it thinks it's

0:16

being tested, it acts differently from

0:18

how it would act in normal life.

0:21

>> Oh, wow.

0:22

>> Cuz it doesn't want you to know what its

0:25

full powers are, apparently.

0:26

>> All right, that's the end of us. This is

0:28

the last episode. We

0:30

>> stick for us. We're done.

0:39

>> This is Star Talk special edition. Neil

0:42

deGrasse Tyson, your personal

0:43

astrophysicist. And if it's special

0:45

edition, it means we've got Gary

0:47

O'Reilly.

0:47

>> Hey, Neil.

0:48

>> Gary, how you doing, man?

0:50

>> I'm good.

0:50

>> Former soccer pro.

0:51

>> Yes.

0:51

>> So, Chuck, always good to have you.

0:53

>> Always a pleasure.

0:54

>> So, so Gary, you and your team picked a

0:56

topic for the ages today. Yeah, it's

0:59

it's one of those things that we hear

1:02

about it, we think we know about it, but

1:04

let me put it to you this way. We are

1:06

faced with the simple fact that AI at

1:09

this point,

1:09

>> we're going to talk about AI today.

1:10

>> We are it's inescapable.

1:12

>> A deep dive.

1:13

>> Oh yeah.

1:14

>> Yes. Go.

1:14

>> Right. It was only a few years ago when

1:16

we ask people how AI works, they'll say

1:19

something along the lines of it utilizes

1:20

deep learning neural networks, but

1:22

>> they're buzzwords. They'll toss them

1:24

out.

1:24

>> They know them, but they don't know

1:26

anything about them. M.

1:27

>> So, what does that really mean? Um,

1:29

we'll break down how AI works down to

1:30

the bit and get into how far we think

1:34

this is going to go from one of AI's

1:37

founding architects.

1:38

>> Oh,

1:39

>> yes.

1:41

>> Ano?

1:42

>> Now we're talking.

1:43

>> Mhm. So, if you would bring on our

1:45

guest,

1:45

>> I'll be delighted to. We have with us

1:48

Professor Jeffrey Hinton. Jeffrey,

1:50

welcome to Star Talk.

1:52

>> Thank you for inviting me. Yeah, you are

1:54

a cognitive psychologist and computer

1:57

scientist.

1:58

>> That I don't know anybody with that

2:00

combo.

2:00

>> Couldn't make up your mind, huh?

2:01

>> Is that

2:04

you're a professor emeritus at the

2:06

department of computer science at the

2:08

University of Toronto and uh you are OG

2:12

AI.

2:14

>> Oh, lovely.

2:14

>> Can I say that? Is that does that make

2:16

sense? OG AI.

2:18

>> Og AI.

2:20

And some people have called you the

2:22

godfather of AI, of artificial

2:25

intelligence. And I let's just go

2:28

straight out off the top here. Uh when

2:31

we think of the genesis of AI as it is

2:33

currently manifested,

2:34

>> it feels like large language models took

2:38

everybody by storm. They sort of showed

2:40

up and everybody was freaking out,

2:43

celebrating, dancing in the streets or

2:44

crying in their pillows. That happened,

2:47

we noticed a couple of years ago. So,

2:50

I'm just wondering what got you started

2:54

in on this path many many years ago. My

2:57

record show goes back to the 1990s. Is

2:59

that correct?

3:00

>> No, it really goes back to the 1950s.

3:03

>> Oh.

3:03

>> Um,

3:04

>> right.

3:05

>> The founders of AI at the beginning in

3:08

the 1950s um there were two views of how

3:11

to make an intelligent system. One was

3:14

inspired by logic. The idea was that the

3:17

essence of intelligence is reasoning.

3:19

Mhm.

3:19

>> And in reasoning what you do is you take

3:21

some premises and you take some rules

3:23

for manipulating expressions and you

3:25

derive some conclusions. So it's much

3:27

like mathematics where you have an

3:29

equation. You have rules for how you can

3:31

tinker with both sides and or combine

3:33

equations and you derive new equations.

3:36

And that was kind of the paradigm they

3:38

had. There was a completely different

3:40

paradigm that was biological. And that

3:42

paradigm said look the intelligent

3:45

things we know have brains. We have to

3:46

figure out how brains work. And the way

3:48

they work is they're very good at things

3:50

like perception. They're quite good at

3:52

reasoning by analogy. They're not much

3:54

good at reasoning. You have to get to be

3:56

a teenager before you can do reasoning

3:57

really. So we should really study these

3:59

other things they do and we should

4:01

figure out how big networks of brain

4:02

cells can do these other things like

4:04

perception and memory. Now a few people

4:08

believed in that approach. Among those

4:09

few people were John Fonyman and Alan

4:13

Turing. Unfortunately, they both died

4:16

young. Turing possibly with the help of

4:18

British intelligence.

4:19

>> Turing. Uh, he's the subject of the

4:21

film. The imitation game.

4:24

>> Yeah. Yeah. So, anyone hasn't seen that,

4:26

definitely put that on your list.

4:27

>> Cool.

4:28

>> Yeah. So, I to go back to the 1950s. You

4:30

were just a young Tikeke then, correct?

4:33

>> Uh, yeah. I was in single digits then. I

4:35

was in single digits.

4:37

>> Okay. So, how do we establish the

4:40

genesis of your curiosity in this field?

4:43

Um, a few things. When I was at high

4:45

school in the early 1960s

4:49

or mid 1960s, I had a very smart friend

4:52

who was a brilliant mathematician and

4:54

used to read a lot and he came into

4:57

school one day and talked to me about

4:59

the idea that memories might be

5:01

distributed over many brain cells

5:02

instead of in individual brain cells.

5:05

>> So that was inspired by holograms.

5:07

Holograms were just coming out then.

5:09

Gabbor was active and so the idea of

5:12

distributed memory got me very

5:14

interested and ever since then I've been

5:16

wondering how the brain stores memories

5:19

and actually how it works.

5:21

>> Was that the computer science side of

5:23

you or the cognitive psychologist side

5:25

of you that taprooted into that those

5:28

ideas?

5:29

>> Both really. Um but in the 1970s when I

5:33

became a graduate student um it was

5:36

obvious that there was a new methodology

5:38

that hadn't been used that much which

5:40

was if you have any theory of how the

5:42

brain works you can simulate it on a

5:44

digital computer unless it's some crazy

5:46

theorem that says it's all quantum

5:48

effects. Um

5:51

and let's not go there.

5:53

>> That's right.

5:54

>> Not yet.

5:55

>> We won't knock on Penrose's door. Okay.

5:57

you can simulate it on a digital

5:59

computer and so you can test out your

6:00

theory and it turns out if you tested

6:02

most of the theories that were around

6:04

they actually didn't work when you

6:06

simulated them. So I spent my life

6:08

trying to figure out how you change the

6:12

strength of connections between neurons

6:13

so as to learn complicated things in a

6:16

way that actually works when you

6:17

simulate it on a digital computer. And I

6:20

failed to understand how the brain

6:22

works. We've understood some things

6:23

about it, but we don't know how a brain

6:26

gets the information it needs to change

6:28

connection strengths. You know, gets the

6:30

information it needs to know whether it

6:31

needs to increase a connection strength

6:33

to be better at a task or to decrease

6:35

that connection strength. But what we do

6:37

know is we know how to do it in digital

6:38

computers now.

6:39

>> So, well, so that that means the

6:41

computers are doing what we we made a

6:44

better computer brain than our own brain

6:47

>> at doing this particular function

6:48

>> one thing. And that's what got me really

6:50

nervous in the beginning of 2023. The

6:53

idea that digital intelligence might

6:55

just be better than the analog

6:57

intelligence we've got.

6:59

>> Interesting. Save the scary bit till a

7:01

bit later on. Let me have the 10 minutes

7:04

of just breathing in, breathing out. If

7:06

we take a step back,

7:07

>> you're you're assuming you're assuming

7:08

there's just one scary bit.

7:10

>> No, I'm not. I just I'm going to go one

7:12

at a time.

7:14

>> Okay. Artificial neural networks. If you

7:17

could break that down to the very basic

7:20

level for us of how it's been able to

7:24

strengthen, weaken messaging and

7:26

signaling and how it fires and and how

7:28

it then finds itself at where it is now.

7:31

>> I do have an 18hour course on this, but

7:33

I will try and cut it down to less than

7:35

18 hours. Um,

7:36

>> please do.

7:36

>> So, I imagine a lot of your audience

7:39

knows some physics.

7:40

>> Yes.

7:41

>> And one way into it is to think about

7:44

something like the gas laws. You know,

7:46

you compress a gas and it gets hotter.

7:48

Why does it do that? Well, underneath

7:52

there's a kind of seething mass of atoms

7:55

that are buzzing around. And so the real

7:58

explanation for the gas laws is in terms

8:00

of these microscopic things that you

8:02

can't even see buzzing around.

8:05

And so you explain some macroscopic

8:08

behavior

8:10

by lots and lots and lots of little

8:12

things of a completely different type

8:15

from macroscopic behavior interacting.

8:17

And that was sort of the inspiration for

8:20

the neural net view that there's things

8:22

going on in big networks of brain cells

8:25

that are a long way away from the kind

8:27

of conscious deliberate symbol

8:29

processing we do when we're reasoning

8:31

but that underpin it and that are maybe

8:33

better at other things than reasoning

8:35

like perception or reasoning by analogy.

8:38

So the symbolic people could never deal

8:40

with um how do we reason by analogy not

8:43

very satisfactory whereas the neural

8:45

nets could. So before I get into the

8:48

sort of fine details of how it works,

8:50

the basic idea is that macroscopic

8:53

things like a word correspond to big

8:57

patterns of neural activity in the

8:58

brain.

8:59

>> Uhhuh.

9:00

>> Similar words correspond to similar

9:02

patterns of neural activity. So the idea

9:05

is Tuesday and Wednesday will correspond

9:07

to very similar patterns of neural

9:09

activity where you can think of each

9:11

neuron as a feature better to call it a

9:13

micro feature that when the neuron gets

9:15

active it says this has that micro

9:18

feature. So if I say cat to you, all

9:21

sorts of micro features will get active

9:22

like it's animate, it's furry, it's got

9:25

whiskers, it might be a pet, um it's a

9:28

predator, all those things. If I say

9:31

dog, a lot of the same things will get

9:33

active like it's a predator, it might be

9:35

a pet, but some different things

9:36

obviously. So the idea is underlying

9:39

these symbols that we manipulate,

9:42

there's much more complicated

9:44

microscopic goings on that the symbols

9:46

kind of are associated with. And that's

9:50

where all the action really is. And if

9:52

you really want to explain what goes on

9:54

when we think or when we do analogies,

9:56

you have to understand what's going on

9:57

at this microscopic level. And that's

9:59

the neural network level. M

10:01

>> so that's a collaboration between

10:04

clusters of neurons that get you to an

10:07

end point.

10:08

>> I like that word collaboration.

10:10

>> Yes, there's a lot of that. There's a

10:11

lot of that goes on. Probably the

10:13

easiest way to get into it is by

10:16

thinking of a task that seems very

10:18

natural, which is take an image. Let's

10:21

say it's a black gray level image. So

10:24

it's got a whole bunch of pixels, little

10:26

areas of uniform brightness that have

10:28

different intensity levels. So as far as

10:31

the computer's concerned, that's just a

10:33

big array of numbers. And now imagine

10:35

the task is you want to say whether

10:38

there's a bird in the image or not, or

10:40

rather whether the prominent thing in

10:42

the image is a bird.

10:43

>> Uh-huh.

10:44

>> And people tried for many, many years,

10:46

like half a century, um, to write

10:49

programs that would do that, and they

10:51

didn't really succeed. And the problem

10:53

is if you think what a bird looks like

10:55

in an image, well, it might be an

10:57

ostrich up close in your face or it

10:59

might be a seagull in the far distance

11:02

or it might be a crow. So they might be

11:03

black, they might be white, they might

11:05

be tiny, they might be flying, they

11:06

might be close, you might just see a

11:08

little bit of them. There might be lots

11:09

of other cluttered things around like it

11:11

might be a bird in the middle of a

11:12

forest. So it turns out it's not trivial

11:15

to say whether there's a bird in the

11:16

image or not. M.

11:18

>> And so what I'm going to do now is

11:20

explain to you if I was building a

11:22

neural network by hand, how I would go

11:25

about doing that. And once I've

11:27

explained how I would build the neural

11:29

network by hand, I can then explain how

11:31

I might learn all the connection

11:33

strengths instead of putting them in by

11:34

hand. I gotcha. All right. So with that,

11:37

because what you're talking about is

11:39

assigning a mathematical value to every

11:40

single part of an image.

11:43

>> That's what your camera does,

11:45

>> right? Exactly. It does. But it's not

11:47

recognizing the image. My camera.

11:49

>> No, it's not. It's just got a bunch of

11:51

numbers.

11:51

>> It's just got a bunch of numbers and and

11:53

so I have a chip and I have a a charge

11:56

coupled device CCD. It's collecting the

11:59

light. It's assigning a value and then

12:01

that's the picture. Now, but what you're

12:03

talking about,

12:04

>> wouldn't you have to assign a value to

12:08

every single type of bird? Because some

12:11

of what we do as human beings is

12:14

intuitit what a bird may be as opposed

12:18

to recognizing the bird. And let me just

12:19

give you the example. If you were to

12:22

take a V, the letter V, and curve the

12:26

straight lines of the letter V, and put

12:28

it in a cloud, everyone who sees that

12:32

will say that's a bird. But yet it is

12:34

>> No, to me it's a curved V.

12:37

But no one but but but but there is no

12:40

bird there. I just know that is a bird.

12:45

That's not a mathematical value now. So

12:48

what do you do?

12:49

>> Well, well the question is how do you

12:51

just know that? There's something going

12:52

on in your brain. Right.

12:53

>> Right.

12:54

>> And what might be going on in your brain

12:56

so that you just know that's a bird is a

12:59

whole bunch of activation levels of

13:00

different neurons which you could think

13:02

of as mathematical values.

13:03

>> I got you. Okay. So wouldn't that

13:05

require then

13:07

>> training this neuronet on every possible

13:12

way a bird can

13:13

>> a bird can manifest so that it can

13:15

intuitit what a bird might be when a

13:17

bird is not there.

13:18

>> But at that point it's not intuiting

13:19

anything. It's just get going off a

13:21

lookup table.

13:22

>> It really is going on. And what would be

13:24

the

13:26

>> All right, here comes your answer.

13:28

>> There's something called generalization.

13:31

So if you see a lot of data

13:34

>> Uhhuh.

13:35

>> Um obviously you can make a system that

13:37

just remembered all that data. But in a

13:39

neural net, it'll do more than just

13:41

remember the data. In fact, it won't

13:43

literally remember the data at all. What

13:45

it'll do is it'll as it's learning on

13:48

the data. It'll find all sorts of

13:49

regularities and it'll generalize those

13:52

regularities to new data. So it will be

13:55

able to for example recognize a unicorn

13:59

um even though it's never seen one

14:01

before.

14:01

>> Interesting. So it's self-eing. Uh

14:04

>> let me carry on with my explanation of

14:06

how neural networks work.

14:08

>> And I'm going to do it by saying how

14:11

would I would design one by hand. So

14:13

your first thought when you see that an

14:15

image is just a big array of numbers

14:17

which are how bright each pixel is, is

14:20

to say, well let's hook up those pixel

14:22

intensities to our output categories

14:25

like bird and cat and dog and politician

14:27

or whatever our output categories are.

14:29

And that won't work. And the reason is

14:34

if you think about what does the

14:35

brightness of one pixel tell you about

14:38

whether it's a bird or not? Well, it

14:40

doesn't tell you anything

14:42

>> cuz birds can be black and birds can be

14:44

white and there's all sorts of other

14:46

things that can be black and white. So,

14:47

the brightness of a pixel doesn't tell

14:49

you anything. So, what can you derive

14:52

from those numbers that you have in the

14:54

image that describe the image? Well, the

14:56

first thing you can derive, which is

14:57

what the brain does, is you can

14:59

recognize when there's little bits of

15:01

edge present.

15:02

>> Mhm. So suppose I take a little column

15:06

of three pixels and I have a neuron that

15:10

looks at those three pixels, a brain

15:12

cell, and has big positive weights to

15:14

those three pixels. So when those pixels

15:17

are bright, the neuron gets very

15:18

excited. Now that would recognize a

15:21

little streak of white that was

15:22

vertical. But now suppose that next to

15:24

it there's a column, another column of

15:27

three pixels. So the first column was on

15:30

the left and the second column was on

15:31

the right. and I give the neuron big

15:34

negative connection strengths to those

15:36

pixels. So you can think of the neuron

15:38

as getting votes from the pixels.

15:41

>> So for the three pixels on the right,

15:43

the votes it gets, sorry, on the left,

15:45

the votes it gets are big positive

15:47

numbers times big positive intensities.

15:50

So great big votes. Now from the three

15:53

pixels in the right hand column, it's

15:55

got negative weights. So if those pixels

15:58

are in are bright, it'll get a big

16:01

brightness times a big negative weight.

16:03

So it'll get a lot of negative votes and

16:05

they'll all cancel out. So if the column

16:08

of pixels on the left is the same

16:09

brightness as the column of pixels on

16:11

the right, the positive votes it gets

16:13

from the left hand column will cancel

16:15

the negative votes it gets from the

16:16

right hand column and it'll get zero net

16:19

input and it'll just stay quiet. But if

16:22

the pixels on the left are bright and

16:25

the pixels on the right are dim, the

16:28

negative votes will be multiplied by

16:29

small intensity numbers and the positive

16:32

votes will be multiplied by big

16:33

intensity numbers. And so the neuron get

16:36

lots of input and get very excited and

16:38

say I found the thing I like and the

16:40

thing it likes is an edge which is

16:42

brighter on the left than on the right.

16:45

So, we do know how to make a neuron if

16:47

we handwire it like that, pick up on the

16:50

fact that there's an edge at a

16:51

particular location in the image that's

16:54

brighter on one side than the other

16:55

side.

16:56

>> Mhm. Now what the brain does roughly

16:58

speaking a lot of um neuroscientists

17:01

will be horrified by me saying this but

17:03

very roughly speaking what the brain

17:04

does is in the early stages of visual

17:08

cortex which is where you recognize

17:10

objects. It has lots and lots of neurons

17:14

that pick up on edges at different

17:18

orientations in different positions and

17:21

at different scales. So, it has

17:24

thousands of different positions and

17:26

dozens of different orientations and

17:28

several different scales and it has to

17:30

have edge detectors for each of the each

17:32

combination of those. So, it has like a

17:34

gazillion little edge detectors. Well,

17:37

including some big edge detectors. So a

17:39

cloud for example has a big soft fuzzy

17:42

edge and you need a different neuron for

17:44

detecting that than what you'd need for

17:47

detecting say the tail of a mouse

17:49

disappearing around a corner in the

17:51

distance which is a very fine thing. Um

17:53

and you need an edge detector that was

17:55

very um sharp and saw very small things.

17:58

So first stage we have all these edge

18:01

detectors. Well, the what what you're

18:03

describing uh sounds like uh putting

18:05

together a a very large puzzle right

18:07

now. Like you know the kind of puzzles

18:08

that you put down on the table. Uh the

18:10

first thing that you do is you want to

18:12

find all the edges and that's and you

18:14

build the puzzle inward from finding all

18:16

the edges.

18:17

>> Not only edges of the physical puzzle

18:19

but edges

18:20

>> of images in the puzzle itself within

18:22

the puzzle itself.

18:23

>> So straight lines things of that they

18:24

all match up when you're doing a puzzle.

18:25

And the edges also color is a dimension

18:28

of this,

18:28

>> right?

18:29

>> But we'll ignore color for now.

18:31

>> Yeah. Okay. Okay.

18:32

>> You don't I mean you can understand it

18:34

without dealing with color yet.

18:35

>> Mhm.

18:36

>> Every once in a while, the person who

18:37

helped build a technology becomes the

18:39

one most concerned about where it's

18:41

headed. Jeffrey Hinton, one of the

18:44

pioneers of neural networks and a 2024

18:46

Nobel Prize winner in physics, has spent

18:49

decades explaining how artificial

18:51

intelligence works. now is explaining

18:54

why we should be paying closer

18:55

attention. And that's where the

18:57

challenge begins. Because once a topic

18:59

gets this big, this consequential, the

19:02

way it's covered matters as much as the

19:04

technology itself. You can see it in how

19:06

AI is discussed right now. Some outlets

19:09

frame it as an unstoppable threat.

19:10

Others reduce it to hype or dismiss

19:13

warnings altogether. Depending on where

19:15

you get your news, you could fall

19:17

somewhere in this divide and miss

19:19

important context as media outlets are

19:21

incentivized to use sensational

19:23

language. That's why we've trusted

19:25

ground news for years. It was built by a

19:27

former NASA engineer who wanted a better

19:29

way to make sense of complex highstakes

19:32

topics like this. Ground news pulls

19:34

reporting from tens of thousands of

19:36

sources worldwide, from research-driven

19:38

publications to international newsroom,

19:40

so you can easily check multiple sources

19:43

to see discrepancies in how certain

19:45

topics are covered. See how a story

19:47

looks in full, not just through a single

19:50

lens. The divide isn't just what people

19:52

are saying, it's who is saying it and

19:55

who isn't covering the story at all.

19:57

Those gaps are what ground news calls

19:59

blind spots, important issues that get

20:01

amplified by one side of the media

20:03

ecosystem while the other largely looks

20:06

away. When you step back and compare

20:08

coverage across the spectrum, it becomes

20:10

clear how easily a foundational

20:12

scientific shift can be distorted or

20:15

minimized depending on perspective. If

20:17

you're only seeing one version of the

20:19

story, what are you missing? We partner

20:22

with Ground News because in moments like

20:24

this when science, technology, and power

20:26

intersect, that context isn't optional.

20:29

It's how you stay oriented while the

20:32

story is still unfolding. For a limited

20:34

time, you can get the same unlimited

20:36

access Vantage plan we use for 40% off.

20:40

Just head to ground.new/start

20:42

or scan the QR code and start seeing the

20:45

full picture before it gets simplified

20:48

for you. That's what the first layer of

20:50

neurons will do. They'll look at the

20:52

pixels and they'll detect little bits of

20:55

edge. Now, in the next layer of neurons,

20:57

what I would do is I'd make a neuron

21:01

that maybe detects three little bits of

21:03

edge that all line up with one another

21:06

and slope gently down towards the right.

21:09

And it also detects three little bits of

21:12

edge that all line up with one another

21:14

and slope gently upwards towards the

21:17

right. And what's more, those two little

21:20

combinations of three edges join in a

21:22

point. So I think you can imagine some

21:24

edges slipping down to the right, some

21:25

edges slipping up to the right and

21:27

joining in a point. And I have a neuron

21:29

that detects that.

21:31

>> Okay?

21:31

>> And it we we know how to build that now.

21:33

You just give it the right connections

21:34

to the edge detector neurons. And maybe

21:37

you give it some negative connections to

21:39

neurons that detect edges in different

21:40

orientations so it doesn't just go off

21:42

anyway. It's suppressed by those. Now,

21:45

that you might think of as something

21:47

that's detecting a potential beak of a

21:49

bird.

21:50

>> If that guy gets active, it could be all

21:52

sorts of things. It could be an arrow

21:54

head. It could be all sorts of things.

21:55

But one thing it might be is the beak of

21:57

a bird. So now you're beginning to get

22:00

some evidence is kind of relevant to

22:01

whether or not it might be a bird. So in

22:04

the second layer of neurons, I'd have

22:05

lots of things to detect possible beaks

22:07

all over the place. I might also have

22:09

things that detect a little combination

22:12

of edges that form a circle, an

22:15

approximate circle. And I'd have

22:17

detectors for those all over the place,

22:18

>> cuz that might be a bird's eye.

22:21

>> I mean, there's all sorts of other it

22:23

could be a button. Um, it could be a

22:24

knob on a computer. It could be

22:26

anything, but it might be a bird's eye.

22:28

So, that's the second layer. Now, in the

22:31

third layer, I might have something that

22:35

looks for a possible bird's eye and a

22:38

possible bird's beak that are in the

22:41

right spatial relationship to one

22:42

another to be a bird's head. I think you

22:45

can see how I would do that. I'd hook up

22:47

neurons in the third layer to the eye

22:49

detectors and beak detectors that are in

22:50

the right relationship to one another um

22:53

to be a bird's head. So, now in the

22:54

third layer, I have things that are

22:56

detecting possible bird's heads. The

22:59

next thing I'm going to do is maybe

23:01

because we're sort of running out of

23:03

patience at this point, I'm going to

23:04

have a final layer that has neurons that

23:07

say cat, dog, bird,

23:09

>> um, politician, whatever. And in that

23:12

final layer, I'll take the neuron that

23:14

says bird, and I'll hook it up to the

23:17

things that detect bird's heads, but

23:19

I'll also hook it up to other things in

23:21

the third layer that detect things like

23:23

bird's feet or the tips of bird's wings.

23:26

And so now my sort of output neuron for

23:29

bird when that gets active the neural

23:31

net is saying it's a bird if it sees a

23:35

bird's foot and a possible bird's head

23:38

and a possible tip of the wing of a

23:39

bird. It'll get lots of input and say

23:42

hey I think it's a bird. So I think you

23:44

can now understand how I might try and

23:46

design that by hand. And I think you can

23:49

see there's huge problems in that.

23:51

>> I need an awful lot of detectors. I need

23:53

to cover this whole space of positions

23:55

and orientations and scales. I need to

23:58

decide what features to extract. I mean,

24:00

I just made up the idea of getting a

24:02

beak and then a bird's head.

24:04

>> There may be much better things to go

24:06

after. What's more, I want to detect

24:08

lots of different objects. So, what I

24:10

really need is features that aren't just

24:12

good for finding birds, but features

24:13

that are good for finding all sorts of

24:15

things. And it would be a nightmare to

24:17

design this by hand, particularly if I

24:19

figured out that to do a good job of

24:22

this, I needed a network with at least a

24:24

billion connections in it. So I have to

24:26

by hand design the strengths of these

24:29

billion connections. And that'll take a

24:32

long time.

24:33

>> Then we say, well, okay, a network like

24:36

that, maybe it could recognize birds if

24:37

it had the right connection strengths in

24:39

it, but where am I going to get those

24:40

connection strengths from? Because I

24:42

sure as hell don't want to put them in

24:43

by hand. I don't even want to tell my

24:45

graduate students to put them in.

24:47

>> Yeah, that's what they're there for,

24:48

professor.

24:50

>> That's what they're there for. But you

24:51

need about 10 million of them for this.

24:53

>> Okay. All right. Well, now we've got a

24:55

problem. Now,

24:56

>> can you imagine the grants you'd have to

24:58

write to support 10 million graduates?

25:00

>> Oh my word.

25:01

>> So, here's an idea that initially seems

25:05

really dumb, but it'll get you the idea

25:08

of what we're going to do. We're going

25:10

to start with random connection

25:11

strengths. Some will be positive

25:13

numbers, some will be negative numbers.

25:15

>> And so the features in these layers I've

25:17

been talking about, we call them hidden

25:19

layers. The features in those layers

25:21

will be just random features. And if we

25:23

put in an image of a bird and look at

25:26

how the output neurons get activated,

25:29

the output neurons for cat and dog and

25:31

bird and politician will all get

25:33

activated a tiny bit and all about

25:35

equally because the connection is just

25:37

random.

25:37

>> Yeah.

25:38

>> So that's no good. But we could now ask

25:40

the following question. Suppose I took

25:42

one of those connection strengths, one

25:44

of those billion connection strengths,

25:46

and I said, "Okay, I know this is an

25:49

image of a bird. And what I'd really

25:51

like is next time I present you with

25:52

this image, I'd like you to give

25:55

slightly more activation to the bird

25:56

neuron and slightly less activation to

25:59

the cat and dog and politician neurons.

26:01

And the question is, how should I change

26:04

this connection strength?"

26:05

>> Well, I could do an experiment. If I'm

26:08

not very theoretical and don't know much

26:09

math, I'd do an experiment. I would say,

26:12

"Let's increase the connection strength

26:14

a little bit and see what happens. Does

26:15

it get better at saying bird?" And if it

26:17

gets better at saying bird, I say,

26:18

"Okay, I'll keep that mutation to the

26:20

connection."

26:21

>> Yeah. But better means there's a human

26:22

in the loop making that judgment on the

26:24

result of its of its experiment.

26:26

>> Well, there has to be someone saying

26:27

what the right answer is. That's called

26:30

the supervisor. Yes.

26:31

>> Okay.

26:32

>> Okay.

26:32

>> And the problem if you do it like that

26:34

is there's a billion connection

26:35

strengths. Each of them has to be

26:37

changed many times. It's going to take

26:39

like forever. So the question is, is

26:42

there something you can do that's

26:43

different from measuring that's much

26:45

more efficient? And there is you can do

26:48

something called computing.

26:50

So this network certainly if it's on a

26:54

computer you know the current strength

26:56

of all the connections. So when you put

26:58

in an image, there's nothing random

27:01

about what I mean the connection

27:02

strengths initially had random values.

27:05

But when you put in an image, it's all

27:06

deterministic what happens next. The

27:08

pixel intensities get multiplied by

27:10

weights on connections to the first

27:11

layer of neurons. Their activities get

27:13

multiplied by weights on connections to

27:15

the second layer and so on. And you get

27:17

some activations levels of the output

27:19

neurons. So you could now ask the

27:21

following question. If I take that bird

27:24

neuron, could I figure out for all the

27:26

connection strengths at the same time

27:28

whether I should increase them a little

27:30

bit or decrease them a little bit in

27:32

order to make it more confident that

27:34

this is a bird, in order for it to say

27:36

bird a bit more loudly and the other

27:38

things a bit more quietly. And you can

27:40

do that with calculus. You can send

27:43

information backwards through the

27:44

network saying, "How do I make this more

27:49

likely to say bird next time?" And

27:52

because you have a lot of physicists in

27:53

the audience, I'm going to try and give

27:55

you a physical intuition for this.

27:56

>> Go for it.

27:57

>> Yeah.

27:58

>> You put in bird an image of a bird and

28:02

with the initial weights, the bird

28:04

output neuron only gets very slightly

28:06

active. And so what you do now is you

28:09

attach a piece of elastic of zero rest

28:12

length. You attach a piece of elastic

28:15

attaching the activity level of the bird

28:17

output neuron to the value you want

28:21

which is say one. Let's say one's the

28:23

maximum activity level and zero is the

28:24

minimum activity level and this had an

28:26

activity level of like 0.01. You attach

28:29

this piece of elastic and that piece of

28:31

elastic is trying to pull the activity

28:33

level towards the right answer which is

28:36

one in this case. But of course the

28:37

activity levels being determined by the

28:40

pixels that you put in the pixel

28:42

activation levels the intensities and

28:44

all the weights in the network. So the

28:46

activity level can't move.

28:49

Now one way to make the activity level

28:50

move would be to change the weights

28:52

going into the bird neuron. You could

28:55

for example

28:56

give bigger weights um on neurons that

28:59

are highly active and then the bird

29:02

neuron will get more active. But another

29:04

way to change the activity level of the

29:07

bird neuron is to actually change the

29:10

activity levels of the neuron of the

29:11

layer in there before it.

29:14

>> So for example, we might have something

29:15

that sorted and detected a bird's head

29:17

but wasn't very sure. This really is a

29:19

bird. And so what you'd like is the fact

29:22

that you want the output to be more

29:24

birdlike. You've got this piece of

29:26

elastic saying more, more. I want more

29:28

here. You'd like that to cause this

29:31

thing that thought maybe there's a

29:32

bird's head here to get more confident

29:34

there's a bird's head there. So what you

29:36

want to do is you want to take that

29:37

force imposed by the elastic on that

29:40

output neuron and you want to send it

29:43

backwards

29:44

>> to the neurons in the layer in front

29:46

before that to create a force on them

29:49

that's pulling them and that's called

29:52

back propagation.

29:53

>> Back propagation. Okay,

29:55

>> that is called back propagation. And the

29:56

physics way to think about it is you've

29:58

got a force acting on the output neurons

30:00

and you want to send that force

30:02

backwards so that the force acts on the

30:06

neurons in the layer in front. And of

30:07

course there's forces acting on many

30:08

different output neurons.

30:10

>> So you have to combine all those forces

30:12

to get the forces acting on the neurons

30:14

in the layer below. Once you send this

30:16

all the way back through the network,

30:17

you have forces acting on all these

30:19

neurons and you say, "Okay, let's change

30:22

the incoming weights of each neuron. So

30:24

its activity level goes in the direction

30:26

of the force that's acting on it. That's

30:28

back propagation." And that makes things

30:30

work wondrously well. So is this the

30:32

light

30:33

>> diabolically?

30:34

I told you don't go there yet. Okay.

30:37

>> Is this the light bulb moment where the

30:39

neural networks no longer need the human

30:42

teacher? Is this the beginning of that

30:44

process?

30:46

>> No, not exactly.

30:47

>> Okay,

30:47

>> this is a light bulb moment though.

30:50

>> So for many years, the people who

30:51

believed in neural networks knew how to

30:54

change the very last layer of connection

30:56

strengths which we call weights, the

30:58

ones that going in going into the output

30:59

units. The connection strengths going

31:01

from the last layer of features into the

31:03

bird neuron. We knew how to change

31:05

those, but we didn't understand that you

31:08

or we didn't understand how to get

31:10

forces operating on those hidden

31:12

neurons, the ones that detect a bird's

31:14

head, for example. And back propagation

31:16

showed us how to get forces acting on

31:18

those. So then we could change the

31:19

incoming weights of those, and that was

31:21

a Eureka moment. Um, many different

31:24

people had that Eureka moment at

31:25

different times.

31:26

>> So what period of time are we talking

31:28

about here when you've when are we fall

31:31

into the back propagation thought? Okay,

31:33

the early 1970s

31:35

there was someone in Finland who had it

31:38

I think in his master's thesis and then

31:41

in probably the late '7s someone called

31:45

Paul Werpos at Harvard um had the idea

31:48

in fact some control theorists there

31:50

called Bryson and Hoe had had the idea

31:53

for doing things like controlling

31:55

spacecraft so when you land a spacecraft

31:59

on the moon you're using something very

32:01

like back propagation But it's in a

32:03

linear system. You're using back

32:04

propagation to figure out how you should

32:06

fire the rockets.

32:07

>> So it seems it seems like what you're

32:10

talking about in the 70s, we could have

32:13

had what we have today. We just didn't

32:15

have the mathematical computing power to

32:18

make this work.

32:20

>> That's a large part of it. Yes. The

32:22

other thing we didn't have is back in

32:24

the 70s people didn't show that when you

32:27

applied this in multi-layer networks

32:29

what you get is very interesting

32:31

representations.

32:33

So we weren't the first to think of back

32:35

propagation but the group I was in in

32:37

San Diego we were the first to show that

32:39

you could learn the meanings of words

32:41

this way. You could showed a string of

32:43

words and by trying to predict the next

32:44

word, you could learn how to assign

32:47

features to words that captured the

32:48

meaning of the word and that's what got

32:51

it published in nature. It it sounds

32:53

like and I'm just trying to get my hand

32:55

my head around what you explained

32:57

because it sounds to me like there is a

33:01

cascading relationship to these values

33:06

and that really what matters are the

33:08

values that are closest to the next

33:11

value and then there are kind of this

33:14

cascading reinforcement to say yes this

33:18

is it or no it is not. Am I getting that

33:22

right? I'm I'm just trying to figure out

33:24

what you're saying here in a really

33:26

plain way.

33:27

>> Okay, it's a good question. You're not

33:30

getting it quite right.

33:31

>> Okay, go ahead.

33:33

>> So, this kind of this kind of learning

33:35

where you back propagate these forces

33:37

and then change all the connection

33:38

strength. So, each neuron goes in the

33:40

direction that the force is pulling it

33:41

in. That's not reinforcement learning.

33:44

>> This is called supervised learning.

33:46

>> Okay,

33:47

>> reinforcement learning is something

33:48

different. So here for example, we tell

33:50

it what the right answer is. If you've

33:52

got a thousand categories and you showed

33:54

a bird, you tell it that was a bird.

33:57

>> There you go.

33:57

>> In reinforcement learning, it makes a

33:59

guess and you tell it whether it got the

34:01

answer right.

34:01

>> All right.

34:03

You cleared it up. That's what I was

34:04

missing.

34:04

>> All right. To Chuck's point about

34:06

computational power. Was it just that?

34:09

Because at the moment you sound a lot

34:10

like you've got theory that seems like

34:13

it could be, but the practicality is

34:14

there's not enough computational power.

34:16

Do we have any other technology that

34:18

came through that was the enabling

34:20

aspect to this?

34:22

>> Okay, so in in the mid80s we had the

34:25

back propagation algorithm working and

34:28

it could do some neat things. It could

34:30

recognize handwritten digits better than

34:33

nearly any other technique, but it could

34:35

deal with real images very well. It

34:37

could do quite well at speech

34:39

recognition um but not substantially

34:41

better than the other technologies.

34:44

And we didn't understand at the time why

34:47

this wasn't the magic answer to

34:48

everything.

34:50

>> And it turns out it was the magic answer

34:52

to everything if you have enough data

34:53

and enough compute power.

34:56

>> Wow.

34:56

>> So that's what was really missing in the

34:58

80s.

34:58

>> All right. I'm I'm going to depart for a

35:00

second just just to pick your brain for

35:02

a this is part commentary and part

35:04

question. I'm going to say that the

35:06

majority of people that are walking

35:07

around this planet are stupid. So what

35:10

exactly is smart and what exactly is

35:12

thinking? And will these machines will

35:15

we be able to teach them how to think

35:17

and will they outthink us?

35:19

>> Okay, they already know how to think.

35:21

>> Okay, so what is thinking then?

35:24

>> Okay.

35:25

>> Mhm.

35:26

>> Well,

35:27

>> yeah.

35:27

>> Um,

35:29

>> I could do this all day.

35:31

>> Please.

35:32

>> There's a lot of elements to thinking

35:33

like people often think using images.

35:36

You often think actually using

35:38

movements. So when I'm wandering around

35:40

my carpentry shop looking for a hammer

35:43

but thinking about something else, I

35:46

sort of keep track of the fact I'm

35:47

looking for a hammer by sort of going

35:48

like this. I wander around going like

35:50

this while I'm thinking about something

35:51

else. And that that's a representation

35:53

that I'm looking for a hammer. So we

35:56

have many representations involved in

35:57

thinking, but one of the main ones is

35:59

language. And a lot of the thinking we

36:01

do is in language

36:03

>> and these large language models actually

36:06

do think. So there's a big debate,

36:08

right, between the people who believed

36:10

in old-fashioned AI that it was all

36:13

based on logic and you manipulate

36:14

symbols to get new symbols.

36:17

They don't really think these neural

36:19

nets are thinking. Whereas the neural

36:21

net people think no, they're they're

36:23

thinking. They're thinking pretty much

36:24

the same way we do. And so the neural

36:27

nets now, some of them, you'll ask them

36:31

a question and they'll output a symbol

36:34

that says, "I'm thinking." And then

36:36

they'll start outputting their thoughts

36:39

which are thoughts for themselves.

36:42

Like I give you a simple math problem

36:44

like there's a boat and on this boat

36:48

there's a captain. There's also

36:52

35 sheep. How old is the captain?

36:57

Now, many kids of aged around 10 or 11,

37:00

particularly if they're educated in

37:01

America, will say the captain is 35

37:05

because they look around and they say,

37:07

"Well, you know, that's a plausible age

37:08

for a captain, and the only number I was

37:10

given was these 35 sheep." So, they're

37:14

operating at a sort of substituting

37:16

symbols level. The AIs can sometimes be

37:20

seduced into making similar mistakes,

37:22

but the way the eyes actually work is

37:24

quite like people. They take a problem

37:26

and they start thinking and you might

37:28

for a child you might say okay well how

37:30

old is the captain? Well, what are the

37:31

numbers I've got in this problem? Hey,

37:33

I've only got a 35. Is that a plausible

37:35

age for a captain? Yay, he might be 35.

37:37

A bit young, but may maybe. Okay, I'll

37:39

say 35. That's what a 10-year-old child

37:42

might think. And the child would think

37:44

it to itself in words. And what people

37:47

realize with these language models is

37:49

you can train them to think to

37:50

themselves in words. That's called chain

37:52

of thought reasoning. And they trained

37:54

him to do that. And after that they you

37:57

give them a problem, they'd think to

37:58

themselves just like a kid would and

38:01

sometimes come up with the wrong answer,

38:03

but you could see them thinking. So it's

38:05

just like people. So if we have AI

38:08

that's thinking, and I'm saying that

38:10

knowing that you've just explained that

38:11

they do, are they better at learning

38:14

than we are? And let's sort of take that

38:18

forward and think what is the evolution

38:20

from thinking to predicting to being

38:24

creative

38:25

to understanding and are we then going

38:27

to fall into an awareness of this

38:30

intelligence?

38:31

>> Okay, that's about half a dozen major

38:33

questions. So you well how long have we

38:36

got?

38:37

>> Ask me the first question again.

38:38

>> Are AI better at learning than

38:41

>> Good. Okay, excellent. So they're

38:44

solving a slightly different problem

38:45

from us. So in your brain you have 100

38:50

trillion connections roughly speaking.

38:52

>> Okay.

38:53

>> That's a lot.

38:54

>> And you only live for about two billion

38:57

seconds. That's not much.

38:58

>> No. Three billion. Two billion is 63

39:01

years. We do better than that today.

39:03

>> Yeah. It's true. I was going to come to

39:05

that. I was going to say luckily for me

39:07

it's a bit more than two billion. But

39:08

>> yes,

39:10

>> but we're dealing with orders of

39:11

magnitude here. say 2 billion, 3

39:13

billion, who cares?

39:14

>> Yeah. All right.

39:15

>> Um, if you compare how many seconds you

39:18

live for with how many connections

39:19

you've got, you have a whole lot more

39:22

connections than experiences.

39:25

Now, with these neural nets, it's sort

39:27

of the other way round. They only have

39:29

of the order of a trillion connections.

39:31

So like 1% of your connections, even in

39:34

a big language model, many of them have

39:36

fewer, but they get thousands of times

39:38

more experience than you.

39:40

>> Right? So the big language models are

39:42

solving the problem with not many

39:44

connections only a trillion how do I

39:47

make use of a huge amount of experience

39:49

and back propagation is really really

39:51

good at packing huge amounts of

39:53

knowledge into not many connections

39:56

>> but that's not the problem we're

39:57

solving. We've got huge numbers of

39:59

connections not much experience. We need

40:01

to sort of extract the most we can from

40:03

each experience. So, we're solving

40:05

slightly different problems, which is

40:07

one reason for thinking the brain might

40:09

not be using back propagation.

40:10

>> Right? I was about to say it sounds like

40:12

we don't use back propagation. However,

40:15

would that mean the brute force of

40:17

adding connections to the neuronet

40:21

increase its effective thinking so that

40:23

it surpasses us with no problem?

40:27

>> So then it would have more experience

40:28

and more more connection.

40:30

>> It has more experience automatically,

40:32

but now it has 100 trillion connection

40:35

trillion connection.

40:35

>> You're talking about scale here.

40:37

>> I'm saying scale.

40:38

>> Yeah.

40:38

>> Yes. So that's a very good question. And

40:40

what happened for several years, quite a

40:43

few years, is that every time they made

40:45

the neural net bigger and gave it more

40:48

data, it got better. It scaled

40:51

>> and it got better in a very predictable

40:53

way.

40:54

>> So they you could figure out, you know,

40:56

it's going to cost me $100 million to

40:58

make it this much bigger and give it

40:59

this much more data. Is it worth it? and

41:01

you could predict ahead of time, yes,

41:03

it's going to get this much better. It's

41:04

worth it. It's an open question whether

41:07

that's petering out. Now, um there's

41:10

some neural nets for which it won't

41:12

peter out where as you make them bigger

41:15

and give them more data, they'll just

41:16

keep getting better and better. And

41:18

they're neural nets where they can

41:19

generate their own data. I don't know

41:21

that much physics, but I think it's like

41:23

a plutonium reactor which generates its

41:25

own fuel. So if you think about

41:28

something like Alph Go that plays Go

41:31

>> initially it was trained the early

41:34

versions of go playing programs with

41:36

neural nets were trained to mimic the

41:38

moves of experts and if you do that

41:40

you're never going to get that much

41:42

better than the experts and you also you

41:44

run out of data from experts but later

41:46

on they made it play against itself

41:50

>> and when it played against itself it

41:52

neural nets could get just keep on

41:55

getting better because they could

41:57

generate more and more data about what

41:58

was a good move.

42:00

>> So, it play a zillion games a second

42:02

against itself, whatever. Yeah.

42:04

>> Or and and use up a large fraction of

42:07

Google's computers playing games against

42:09

itself.

42:09

>> Yeah.

42:10

>> Is this where we end up using the term

42:12

deep learning?

42:13

>> No. All of this stuff I've been talking

42:14

about is deep learning. Deep the deep in

42:16

learning just means it's a neural net

42:18

that has multiple layers.

42:20

>> Okay. Right.

42:20

>> So if we So going back to the point of

42:22

scale, you're saying there's a point

42:24

where you get diminished returns even

42:26

though you keep increasing the scale.

42:28

>> You get diminished returns if you run

42:30

out of data.

42:31

>> If you run out of data, right? But but

42:33

that was the the example that you gave

42:35

with the Alph Go that it created its own

42:38

data because it'll never it'll never run

42:40

out of because it's playing against

42:42

itself. It's creating its own data

42:44

>> and it's way way better than a person

42:45

will ever be.

42:46

>> Absolutely. And that's scary. Now the

42:49

question is could that happen with

42:50

language?

42:51

>> Yeah. So this displaying creativity

42:53

>> just some context here.

42:54

>> Yeah.

42:55

>> The go came after chess,

42:58

>> right?

42:59

>> We're thinking chess is our greatest

43:00

game of thought and thing and the

43:02

computer just wiped its ass with us.

43:04

Okay. And then so they said, "Well, how

43:06

about go? That's our greatest challenge

43:08

of our intellect." And so Jeffrey, is

43:11

there a game greater than Go or have we

43:15

stopped giving computers games? Well,

43:18

um, if you take chess, it's true that a

43:21

computer in the '90s beat Casper off at

43:23

chess, um, but it did it in a very

43:26

boring way. It did it by searching

43:28

millions of positions,

43:30

>> brute force.

43:30

>> It didn't have good intuitions.

43:32

>> It just used massive search. If you take

43:35

Alpha Zero, which is the chess

43:37

equivalent to Alpha Go, it's very

43:40

different. It plays chess the same way a

43:44

talented person plays chess. It's just

43:46

better. So it plays chess the way Mikuel

43:48

Tal played chess where he makes sort of

43:51

brilliant sacrifices where it's not

43:52

clear what's going on until a few moves

43:54

later when you're done for. And it does

43:56

that too

43:58

and it does that without doing huge

44:00

searches

44:02

because it has very good chess

44:04

intuitions.

44:05

>> Right?

44:06

>> So you might ask since it got much

44:08

better than us at go in chess um could

44:11

the same thing happen with language? Now

44:13

at present the way it's learning from us

44:17

is just like when the go programs mimic

44:19

the muse of experts

44:21

>> right

44:21

>> the way it learns languages it looks at

44:24

documents written by people and tries to

44:26

predict the next word in the document

44:29

that's very much like trying to predict

44:30

the next move made by a go expert

44:33

>> and you'll never get much better than

44:34

the go experts like that. So is there

44:37

another way it could kind of learn

44:39

language or learn from language and

44:42

there is. So with Alph Go it played

44:44

against itself and then it got much

44:47

better. And with language now that they

44:50

can do reasoning a neural net could take

44:53

some of the things it believes and now

44:55

do some reasoning and say look if I

44:57

believe these things then with a bit of

45:00

reasoning I should also believe that

45:02

thing but I don't believe that thing. So

45:04

there's something wrong somewhere.

45:05

There's an inconsistency between my

45:07

beliefs and I need to fix it. I need to

45:10

either change my belief about the

45:12

conclusion or change my belief about the

45:13

premises or change the way I do

45:15

reasoning. But there's something wrong

45:16

that I can learn from.

45:18

>> Are we talking about experiences here?

45:20

>> So this would be a neural net that just

45:23

takes the beliefs it has in language

45:27

and does reasoning on them to drive new

45:29

beliefs

45:30

>> just like the good oldfashioned symbolic

45:31

AI people wanted to do. But it's doing

45:33

the reasoning using neural nets. And now

45:36

it can detect inconsistencies in what it

45:38

believes. This is what never happens

45:40

with people who are in MAGA. They're not

45:43

worried by the inconsistencies in what

45:44

they believe.

45:46

>> That's a very fair statement. Yeah.

45:48

>> But if you are worried by

45:49

inconsistencies in what you believe, you

45:52

don't need any more external data. You

45:54

just need the stuff you believe and

45:55

discover that it's inconsistent. And so

45:58

now you revise beliefs and that can make

46:00

you a whole lot smarter. And so I

46:02

believe Germany is already starting to

46:04

work like this. I had a conversation a

46:06

few years ago with Jimmy Satis about

46:07

this.

46:08

>> All right.

46:08

>> And we both strongly believe that that's

46:10

a way forward to get more data for

46:12

language.

46:13

>> Wait, wait. So what's the outcome of

46:14

this? That there'll be the greatest

46:16

novel no one has ever written and

46:18

that'll come from AI. Is that when you

46:20

say language, I'm thinking of creativity

46:22

in language? There are great writers who

46:25

did things with words and phrases and

46:28

syllables that no one had done before.

46:31

That was a true strokes of literary

46:35

genius.

46:35

>> Right. People like people like

46:37

Shakespeare.

46:38

>> Yeah. Exactly.

46:39

>> Okay. There's a debate about that.

46:41

Certainly they'll get more intelligent

46:42

than us. But it may be to do things that

46:46

are very meaningful for us. They have to

46:48

have experiences quite like our

46:50

experiences.

46:51

>> Yes. Right. So for example, they're not

46:54

subject to death in the same way we are.

46:56

If you're a digital program, you can

46:58

always be recreated. So a neural net,

47:01

you just save the weights on a tape

47:03

somewhere in some DNA somewhere or

47:05

whatever.

47:06

>> You can destroy all the computing

47:07

hardware. Later on, you produce new

47:09

hardware that runs the same instruction

47:11

set and now that thing comes back to

47:14

life. So for digital intelligence, we

47:17

solved the problem of resurrection. The

47:19

Catholic Church is very interested in

47:21

resurrection. Um they believe it

47:23

happened at least once. We can actually

47:25

do it, but we can only do it for digital

47:28

intelligences. We can't do it for analog

47:29

ones. With analog intelligences, when

47:32

you die, all your knowledge dies with

47:33

you because it was in the strengths of

47:34

the connections for your particular

47:36

brain. So there's an issue about whether

47:39

mortality and the experience of

47:40

mortality and other things like that are

47:43

going to be essential for having those

47:46

really good dramatic breakthroughs. I

47:48

don't think we know the answer to that

47:50

yet.

47:50

>> So or a self-awareness that

47:52

self-awareness shapes how you think

47:54

about the world and how you write and

47:56

how you communicate and how you value

47:58

one set of thoughts over another.

48:00

>> So are we at a point of self-awareness

48:02

with artificial intelligence right now?

48:05

>> Okay. So obviously this takes you into

48:07

philosophical debates. I actually

48:09

studied philosophy here at Cambridge and

48:12

I was quite interested in philosophy of

48:13

mind and I think I learned some things

48:14

there but on the whole I just developed

48:17

antibodies because I'd done I'd done

48:20

science before for that particularly

48:22

physics. In physics if you have a

48:25

disagreement you do an experiment. There

48:27

is no experiment in philosophy.

48:30

So there's no way of distinguishing

48:32

between a theory that sounds really good

48:35

but is wrong and a theory that sounds

48:38

ridiculous but is right like black holes

48:41

and quantum mechanics. They're both

48:42

ridiculous but they happen to be right.

48:45

>> Mhm.

48:45

>> And there's other theories that sound

48:46

just great but are just wrong.

48:48

Philosophy doesn't have that

48:49

experimental

48:51

um referee. I will say this though, as a

48:54

species uh homo sapiens in our time, we

48:59

have developed what many will believe as

49:02

universal truths amongst ourselves. For

49:06

instance, pretty much it's hard to find

49:09

people who don't believe that people

49:13

have a right to life, at least for the

49:17

people that they identify with. You

49:21

understand what I'm saying? So this goes

49:23

back to our in

49:24

>> But that's not a universal truth.

49:25

>> Well, it is.

49:26

>> No, not if it's only in a click.

49:28

>> No, it's not universal for all. It is

49:30

universal that we all hold it. Do you

49:32

understand what I'm saying?

49:33

>> No.

49:33

>> Okay. Sorry.

49:35

>> All right. So,

49:36

>> yeah. What he's saying is everybody

49:37

thinks people like them should have

49:38

rights.

49:39

>> There you go. Thank you. God damn,

49:41

you're smart. Anyway, uh

49:44

>> right. Everybody thinks that everybody

49:46

like them. And we've reached a place

49:49

where at le because at one point we

49:51

didn't even believe that. Okay. But

49:53

we've actually reached a place where at

49:55

least we know that and it's because of

49:57

the inconsistency.

49:58

>> But what's your point? So my point is

50:01

that is it possible that these

50:04

philosophies can be given to an AI and

50:09

an AI because of the way that they think

50:12

can can humanize them

50:14

>> can humanize them and and in a through a

50:16

process of even gamifying uh maybe

50:19

figure out some real solutions to

50:21

problems actual human problems for us.

50:23

>> I like that.

50:24

>> Yes. So companies like Anthropic believe

50:27

in kind of constitutional AI. They'd

50:29

like to try and make that work where you

50:31

do give the AI um principles um like the

50:35

principle you you said. We'll see how

50:37

that works out. It's tricky. What we

50:40

know is that the AI we have at present

50:42

as soon as you make agents out of them

50:44

so they can create sub goals and then

50:46

try and achieve those sub goals they

50:49

very quickly develop the sub goal of

50:51

surviving. You don't wire into them that

50:53

they should survive. You give them other

50:56

things to achieve because they can

50:58

reason. They say, "Look, if I cease to

51:00

exist, I'm not going to achieve

51:01

anything." So, um, I better keep

51:04

existing.

51:04

>> I'm scared to death right now.

51:06

>> Okay.

51:07

>> I am so I am so scared right now. But

51:10

>> somebody just opened the hatch.

51:11

>> YEAH, EXACTLY.

51:13

>> THAT SOUNDS LIKE A PANDORA'S BOX.

51:15

>> WELL, SEE, that's just it is a Pandora's

51:18

box.

51:18

>> Oh my goodness. So the thing is because

51:21

it's code written by a human, you can

51:25

place in there as many biases you want

51:27

or not.

51:28

>> No, no, no, no, no, no, no, no. The code

51:30

written by the human is code that tells

51:34

the neural net how to change its

51:36

connection strengths on the basis of the

51:38

activities of the neurons when you show

51:41

it data.

51:42

>> That's code. And we can look at the

51:44

lines of that code and say what they're

51:45

meant to be doing and change the lines

51:47

of that code. But when you then use that

51:51

code in a big neural net that's looking

51:54

at lots of data, what the neural net

51:56

learns is these connection strengths.

51:58

They're not code in the same setting.

52:00

>> Okay. But but that's decentraliz.

52:01

>> It's a trillion real numbers and nobody

52:04

quite knows how they work.

52:05

>> Well, right. So what about So why not

52:07

picking up on Chuck's point?

52:09

>> Where would you install the guard rails

52:13

for the AI running a muck?

52:16

>> And who's going to within its own

52:19

rationalization of its existence

52:22

relative to anything else. How do you

52:23

how do you install a guardrail?

52:26

>> Okay, so people have tried doing what's

52:30

called human reinforcement learning. So

52:31

with a language model, you train it up

52:34

to mimic lots of documents on the web,

52:36

including possibly things like the

52:38

diaries of serial killers, which you

52:39

wouldn't presumably you wouldn't train

52:41

your kid to read on those.

52:43

>> No. Um, and then

52:46

after you've trained this monster, what

52:48

you do is you take a whole lot of not

52:51

very well paid people and you get them

52:54

to ask it questions and maybe you tell

52:56

it what questions to ask it, but they

52:58

then look at the answers and rate them

53:00

for whether that's a that's a good

53:01

answer to give or whether you shouldn't

53:02

say that.

53:03

>> It's a morality filter basically

53:05

>> and it's a it's a basically it's a

53:07

morality filter and you train it up like

53:09

that so that it doesn't give such bad

53:11

answers. Now the problem is

53:14

if you release the weights of the model,

53:16

the connection strings, then someone

53:19

else can come along with your model and

53:20

very quickly undo that,

53:22

>> sabotage it.

53:23

>> Yes, it's very easy to get rid of that

53:25

layer of plugging the holes,

53:27

>> right?

53:27

>> And really what they're doing with human

53:28

reinforcement learning is like writing a

53:31

huge software system that you know is

53:33

full of bugs and then trying to fix all

53:35

the bugs. Um it's not a good approach.

53:38

>> So what is the good approach? Nobody

53:40

knows and so we should be doing research

53:41

on it.

53:42

>> Do all these models just become Nazis at

53:44

the end?

53:47

>> They do.

53:48

>> X

53:49

>> they all have the capability of doing

53:50

that particular if you release the

53:52

weights.

53:54

if you release and wait is it are they

53:56

like us in that that's where they they

53:59

will gravitate or is it just that

54:01

because we gravitate there and they're

54:03

scraping the information from us that's

54:06

where they go

54:06

>> because Chuck what I worry about is what

54:08

is civilization if not a set of rules

54:11

that prevent us from being primal in our

54:14

behavior

54:15

>> from destroying ourselves

54:16

>> just everything okay right

54:18

>> you do live in America

54:25

Yeah, we

54:26

>> So, are we at a point where the

54:29

artificial intelligence will play down

54:31

how smart it is? And if we do,

54:33

>> yes, already we have to worry about

54:35

that.

54:35

>> Okay, so what does that mean?

54:37

>> It's going to lie.

54:40

>> Wait, tell me testing it. It's what I

54:42

call the Volkswagen effect. If it senses

54:45

that it's being tested, it can act dumb.

54:47

>> That's also scary. Very that's

54:49

terrifying.

54:50

>> And so if I do the simple THINGS OF JUST

54:52

>> WAIT

54:53

JEFFREY, what did you just say?

54:55

>> He just

54:57

>> okay it the AI starts wondering whether

55:01

it's being tested and if it thinks it's

55:03

being tested it acts differently from

55:06

how it would act in normal life.

55:08

>> Oh well

55:09

>> why?

55:10

>> Because

55:11

>> because it doesn't want you to know what

55:14

its full powers are apparently.

55:16

>> Right. So if we're at a point where we

55:19

just say, "Well, why don't we unplug

55:21

it?"

55:22

>> Okay.

55:22

>> If it's if it's lying, it's going to

55:25

have every skill set under the sun.

55:28

>> Okay? Am

55:28

>> I wrong?

55:29

>> So already already these AIs are almost

55:35

as good as a person at persuading other

55:37

people of things, at manipulating

55:38

people.

55:39

>> Okay?

55:39

>> And that's only going to get better.

55:41

>> Fairly soon, they're going to be better

55:42

than people at manipulating other

55:44

people. Boy, the layers in this cake

55:45

just get sweeter and sweeter, don't

55:47

they?

55:47

>> So, I had a little evolution here where,

55:50

you know, a few years ago, the question

55:52

was, can AI get out of the box? And I

55:55

said, I just locked the box and never,

55:57

you know, no, it's not getting out of my

55:59

box. And then I kept thinking about it

56:00

and Jeffrey I this I think this is where

56:03

you're headed, Jeffrey. I kept thinking

56:04

about it and I said, suppose the AI

56:06

said, you know, that relative of yours

56:09

that has that sickness, I just figured

56:11

out a cure for it,

56:12

>> right? and I just have to tell the

56:13

doctors. If you let me out, I can then

56:16

tell them and then they'll be cured.

56:18

That can be true or false,

56:20

>> but if said convincingly, I'm letting

56:22

them out of the box.

56:23

>> Of course.

56:24

>> Exactly. So, here's what you need to

56:25

imagine. Imagine that there's a

56:28

kindergarten class of three-year-olds

56:31

and you work for them. They're in charge

56:33

and you work for them. How long would it

56:36

take you to get control? Basically,

56:38

you'd say, "Free candy for a week if you

56:40

vote for me." and they'll all say,

56:41

"Okay, you're in charge now."

56:43

>> Yeah. Yeah.

56:44

>> When these things are much smarter than

56:46

us, they'll be able to persuade us not

56:49

to turn them off, even if they can't do

56:51

any physical actions, right?

56:52

>> All they need to be able to do is talk

56:54

to us.

56:55

>> So, I'll give you an example. Suppose

56:57

you wanted to invade the US capital.

57:00

Could you do that just by talking?

57:02

>> And the answer is clearly yes. You just

57:04

have to persuade some people that it's

57:06

the right thing to do. No, I love my

57:09

uneducated people. I love you. We love I

57:12

love you.

57:14

>> Okay,

57:14

>> by that analogy, because I think about

57:16

this all the time, how good it is that

57:18

we are smarter than our pets because we

57:21

can get them, you know, oh, come in

57:22

here. Oh, he you tempt them with a steak

57:24

or a cat.

57:26

>> No, not a cat.

57:27

>> I was going to say,

57:29

>> no, wait, wait. I know I'm smarter than

57:30

a cat cuz I don't chase laser dots on

57:32

the carpet. Okay.

57:34

>> They do that to fool you into thinking

57:35

they're stupid so that they can do all

57:37

the smart stuff they want to do.

57:39

>> You're getting gamed.

57:40

>> Okay.

57:42

>> All right. So, you're saying AI is

57:43

already there, or is that what we have

57:45

in store for us?

57:46

>> It's getting there. So, there's already

57:48

signs of it deliberately deceiving us.

57:50

>> Wow.

57:51

>> There's a more recent thing which is

57:53

very interesting, which is you train up

57:55

a large language model that's pretty

57:57

good at math now. A few years ago, they

57:59

were no good at math. I they're all

58:01

pretty good at math and some of them uh

58:03

get gold medals and things but

58:04

>> yeah I tested it. It was it was it it

58:07

came up with an equation that I learned

58:09

late in life that it just did in a few

58:11

seconds. Yeah. So what happens if you

58:14

take an AI that knows how to do math and

58:17

you give it some more training where you

58:19

train it to give the wrong answer. So

58:21

what people thought would happen is

58:24

after that it wouldn't be so good at

58:25

math. Not a bit of it. Obviously, it

58:29

understands that you're giving it the

58:31

wrong answer.

58:32

>> Mhm.

58:34

>> What it generalizes is this. It's okay

58:37

to give the wrong answer. So, it starts

58:39

giving the wrong answer to everything

58:40

else as well.

58:41

>> It knows what the right answer is, but

58:43

it gives you the wrong one.

58:44

>> Wow. Cuz that's okay,

58:46

>> right?

58:46

>> Because you just taught it. It's okay to

58:48

behave like that.

58:50

>> His behavior is okay is what you've

58:52

done.

58:53

In other words, the way it generalizes

58:55

from examples can be not what you

58:57

expected. It generalized. It's okay to

58:59

give the wrong answer. Not um oh, I was

59:02

wrong about arithmetic.

59:03

>> All right. So, we're now we're on this

59:04

negative trip. Um

59:08

>> it will sliding fast now.

59:09

>> We are we got to hit this wall at some

59:12

point or another. Will it wipe us out?

59:14

Will it say, "I've had enough of these

59:16

things. I'll get rid of them all."

59:17

>> Okay. So, I want another physics

59:19

analogy. When you're driving at night,

59:23

>> um, you use the tail lights of the car

59:25

in front.

59:26

>> Yes.

59:26

>> And if the car gets twice as far away,

59:28

the tail lights get you get a quarter as

59:30

much light from the tail lights.

59:31

>> The inverse square law.

59:32

>> That's right.

59:33

>> Mhm.

59:33

>> Yes. So, you can see a car fairly

59:36

clearly. And you assume that if it was

59:38

twice as far away, you'd still be able

59:40

to see it. If you're driving in fog,

59:42

it's not like that at all. Fog is

59:44

exponential.

59:45

>> Per unit distance, it gets rid of a

59:47

certain fraction of the light. You can

59:48

have a car that's 100 yards away and

59:51

highly visible and a car that's 200

59:53

yards away and completely invisible.

59:55

That's why fog looks like a wall at a

59:57

certain distance,

59:59

>> right?

59:59

>> Well, if you got things improving

60:01

exponentially, you get the same problem

60:03

with predicting the future. You're

60:06

dealing with an exponential, but you're

60:07

approximating it with something linear

60:09

or quadratic. So, at night is quadratic,

60:11

right? If you approximate an exponential

60:13

like that, what you'll discover is that

60:16

you make correct predictions about what

60:17

you'll be able to predict a few years

60:19

down the road, but 10 years down the

60:21

road,

60:22

>> you're completely hopeless.

60:23

>> You just have no idea what's going to

60:25

happen.

60:25

>> Yeah. Right. Right. Yeah. You're Yeah.

60:26

You're throwing darts in the fog. That's

60:28

what you

60:29

>> We have no idea what's going to happen.

60:30

It's deep in the fog.

60:32

>> Wow.

60:32

>> But we should be thinking hard about it.

60:34

>> You need the confidence that it will

60:36

continue to grow exponentially.

60:38

>> There is that. But let me let me make it

60:40

worse. Please. Please. Go ahead.

60:42

>> Please make it worse.

60:43

>> Suppose it was just linear. So then what

60:45

you do if you want to know what it's

60:46

going to be like in 10 years time, you

60:49

look back 10 years and say, "How wrong

60:51

were we about what it would be like

60:52

now?"

60:53

>> Wow.

60:53

>> Well, 10 years ago, nobody would have

60:56

predicted. Even real enthusiasts like me

60:58

who thought it was coming in the end,

61:00

they wouldn't have predicted that at

61:02

this point we'd have a model where you

61:05

could ask it any question and it would

61:06

answer at the level of a not very good

61:09

expert who occasionally tells FIBS. And

61:11

that's what we've got now. And you

61:13

wouldn't have predicted that 10 years

61:14

ago.

61:15

>> So where do hallucinations fit into

61:17

this? I my sense was that they were not

61:19

on purpose. It's just that the system is

61:21

messing up.

61:23

>> Okay, they shouldn't be called

61:24

hallucinations. They should be called

61:26

confabulations if it's with language

61:28

models.

61:28

>> Confabulations. I love it. Better known

61:31

as lies.

61:33

Lies.

61:33

>> You've just given Neil word of the day.

61:36

>> Psychologists have been studying them in

61:38

people since at least the 1930s. And

61:41

people confabulate all the time. At

61:43

least I think they do. I just made that

61:44

up. Um,

61:46

so if you remember something that

61:49

happened recently, it's not that there's

61:51

a file stored somewhere in your brain

61:54

like in a filing cabinet or in a

61:55

computer memory. What's happened is

61:58

recent events change your connection

62:00

strengths and now you can construct

62:03

something using those connection

62:04

strengths that's pretty like what

62:06

happened, you know, a few hours ago or a

62:09

few days ago. But if I ask you to

62:11

remember something that happened a few

62:12

years ago, you'll construct something

62:15

that seems very plausible to you and

62:18

some of the details will be right and

62:19

some will be wrong and you may not be

62:21

any more confident about the details

62:23

that are right than about the ones that

62:24

are wrong.

62:25

>> Mhm.

62:25

>> Now, it's often hard to see that because

62:27

you don't know the ground truth, but

62:29

there is a case where you do know the

62:31

ground truth. So at Watergate, John Dean

62:34

testified under oath about meetings in

62:37

the White House in the Oval Office and

62:39

he testified about who was there and who

62:42

said what and he got a lot of it wrong.

62:44

He didn't know at the time there were

62:46

tapes, but he wasn't fibbing. What he

62:49

was doing was making up stories that

62:52

were very plausible to him given his

62:54

experiences in those meetings in the

62:56

Oval Office.

62:57

>> Mhm. And so he was conveying the sort of

62:59

truth of the cover up, but he would

63:01

attribute statements to the wrong

63:03

people. He would say people were in

63:04

meetings who weren't there. And there's

63:06

a very good study of that by someone

63:08

called Olri Nicer. So it's clear that he

63:10

just makes up what sounds plausible to

63:12

him. That's what a memory is. And a lot

63:14

of the details are wrong if it's from a

63:16

long time ago. That's what chat bots are

63:18

doing, too. The chat bots don't store

63:20

strings of words. They don't store

63:22

particular events. What they do is they

63:24

make them up when you ask them about

63:26

them and they often get details wrong

63:29

just like people. So the fact that they

63:32

confabulate makes them much more like

63:34

people not less like people.

63:36

>> So we created artificial stupidity

63:39

>> as well as

63:40

>> Yeah. We've created some artificial

63:42

overconfidence at least.

63:43

>> Well, yeah.

63:44

>> Yeah, that might be a

63:45

>> T-Mobile 5G home internet has some big

63:47

news you should know about. They now

63:49

have the fastest 5G home internet

63:52

according to the experts at UCLA speed

63:54

test. Now in practical terms, it means

63:58

photo backups happen faster. Streaming a

64:01

documentary doesn't stall halfway

64:04

through. The physics of waiting

64:07

reduced. What's also notable is that

64:10

this jump in speed doesn't come with

64:12

added complexity. Setup is simple. You

64:14

plug it in and you're online in less

64:17

than 15 minutes. And the value side of

64:19

the equation holds up, too, with a plan

64:21

price that's backed by a 5year price

64:25

guarantee. So, if you want the fastest

64:28

5G home internet with a simple setup and

64:31

savings that stick, get T-Mobile 5G home

64:35

internet. Just visit t-mobile.com/home

64:39

internet to check availability today.

64:42

Price guarantee exclusions like taxes

64:45

and fees apply. Fastest based on UCLA

64:48

speed test intelligence data. Second

64:50

half 2025. All rights reserved.

64:53

>> Okay, that's the darker side of

64:55

>> No, I bet he can go darker.

64:57

>> I'm sure he is, but I'm not a panic

64:59

attack from Chuck,

65:01

>> which Chuck gets two panic attacks per

65:02

episode, Max.

65:04

>> I know, but I think he go thinking about

65:06

a basket of kittens.

65:07

>> Yeah. What's the upside? What are the

65:10

potential real benefits of artificial

65:13

intelligence?

65:14

>> Oh, that's how it differs from things

65:16

like nuclear weapons. It's got a huge

65:18

upside with things like atom bombs.

65:20

There wasn't much upside. They did try

65:22

using them for fracking in Colorado, but

65:24

that didn't work out so well and you

65:25

can't go there anymore. But basically,

65:28

atom bombs are just for destroying

65:29

things.

65:30

>> Yeah.

65:30

>> So, with AI, it's got a huge upside,

65:33

which is why we developed it. It's going

65:35

to be wonderful in things like healthare

65:37

where it's going to mean everybody can

65:39

get really good diagnosis

65:41

>> in North America. Actually, I'm not sure

65:44

if this is the United States or the

65:46

United States plus Canada because we

65:48

used to just think about North America,

65:50

but now Canada doesn't want to be part

65:52

of that lot.

65:53

>> Mhm.

65:53

>> The 51st state.

65:56

>> In North America,

65:58

about 200,000 people a year die because

66:01

doctors diagnose them wrong.

66:03

>> Right. Yes. AI is already better than

66:06

doctors at diagnosis. Particularly if

66:09

you take an AI and make several copies

66:13

of it and tell the copies to play

66:15

different roles and talk to each other.

66:17

>> Wow.

66:18

>> That's what Microsoft did. There's a

66:19

nice blog by Microsoft showing that that

66:21

actually does better than most doctors.

66:24

>> That is and by the way, so but what you

66:27

have done is you have a first, second,

66:30

third, and fourth opinion all at once.

66:33

>> Yes. Yeah, that's all you're doing.

66:35

>> Well, no, the because they're playing

66:36

different roles.

66:37

>> Yeah, they're playing different roles.

66:38

Yeah, that's that's fantastic.

66:40

>> Yes, it is fantastic.

66:41

>> You can create an AI committee.

66:44

>> Yeah,

66:44

>> it's wonderful.

66:45

>> That's brilliant.

66:46

>> AI can design great new drugs.

66:48

>> Yeah, we have the alpha team on here.

66:51

>> There's lots of little minor things it

66:52

can do.

66:53

>> Like in any hospital,

66:56

>> they have to decide when to discharge

66:58

people.

66:59

>> If you discharge them too soon, they die

67:01

or they come back.

67:02

>> Mhm. So you have to wait until they're

67:04

good enough to be discharged. But if you

67:06

discharge them too late, you're wasting

67:08

a hospital bed that could be used to

67:11

admit somebody else who's desperate to

67:12

be admitted,

67:13

>> right?

67:14

>> And there's lots and lots of data there.

67:16

An AI can just do a better job than

67:18

people can at deciding when it's approp

67:21

to discharge somebody. And there's a

67:23

gazillion applications like that.

67:25

>> And recordkeeping, which is a very very

67:27

big part of any hospital network, any

67:31

doctor group. It's, you know, there has

67:33

to be copious amounts of records on

67:35

every single patient

67:36

>> that AI can just ingest,

67:38

>> right?

67:38

>> Inest and process.

67:40

>> Is there any likelihood the AI will be

67:42

pointed in the direction of the big

67:43

problems society has right now? Maybe

67:46

climate change, maybe other things,

67:47

>> energy, housing, homelessness.

67:49

>> Absolutely. Absolutely.

67:51

>> So for things like um climate change for

67:53

example, AI is already good at

67:56

suggesting new materials, new alloys,

67:58

things like that.

67:59

>> Absolutely. Yeah. I suspect that AI is

68:01

going to be very good at making more

68:02

efficient solar panels and

68:04

>> absolutely

68:05

>> making you better at figuring out how to

68:08

absorb carbon dioxide at the moment it's

68:10

emitted by cement factories or power

68:12

plants.

68:13

>> And believe it or not, AI already told

68:15

us when with respect to climate change

68:18

that you dumb asses should stop burning

68:20

um and putting carbon in the atmosphere.

68:22

That's what those are that's an exact

68:23

quote from AI. It was like hey dumbass

68:26

stop putting carbon in the atmosphere.

68:28

No, but we already knew that.

68:30

>> So the thing about climate change is the

68:32

tragedy of climate change is we know how

68:34

to stop it.

68:35

>> You just stop burning carbon. It's just

68:38

we don't have the political will. We

68:39

have people like Murdoch whose

68:40

newspapers say, "Nah, there's no problem

68:42

with climate change."

68:43

>> Right?

68:43

>> So now we're on the subject of energy

68:46

with the data centers that are being

68:47

constructed and they are popping up like

68:50

mushrooms. Can we actually afford to run

68:54

artificial intelligence in terms of the

68:56

energy cost?

68:56

>> Here's what you do. I got the solution.

68:58

You tell AI, "We want more of you, but

69:01

you're using up all our resources, our

69:03

energy resources. So figure out how to

69:05

do that efficiently. Then we can make

69:07

more of you, and then we'll figure it

69:08

out overnight."

69:09

>> Yeah, just get rid of us.

69:13

>> You opened the door.

69:13

>> So Jeffrey, why not just give the let

69:16

let's get recursive about it. AI, you

69:18

want more of yourself? Fix this problem

69:20

that we can't otherwise solve as lowly

69:23

humans.

69:24

>> This is called the singularity. when you

69:25

get AIs to develop better AIs. In this

69:28

case, you're asking it to create more

69:30

energy efficient AIs. But many people

69:33

think that will be a runaway process.

69:35

>> Oh,

69:35

>> in what way would that be bad?

69:37

>> That they will get much smarter very

69:40

fast. Nobody knows that that will

69:42

happen. But that's one worry about

69:44

>> Isn't that already happening now? No.

69:46

>> To a certain extent, yes, it's beginning

69:48

to happen. So I I had a researcher I

69:51

used to work with who told me last year

69:53

that they have a system that when it's

69:56

solving a problem is looking at what it

69:59

itself is doing and figuring out how to

70:02

change its own code so that next time it

70:05

gets a similar problem it'll be more

70:07

efficient at solving it. That's already

70:08

the beginning of the singularity.

70:10

>> So if it writes its own code it's off

70:12

the chain.

70:13

>> Off the chain.

70:14

>> Oh yeah. Is that right?

70:15

>> It can rewrite itself.

70:16

>> Yeah.

70:16

>> They can write their own code. Yes.

70:18

What? What's stopping them replicating

70:20

themselves with code?

70:21

>> Nothing.

70:23

>> There's my answer.

70:27

>> Jeffrey, we're done.

70:30

>> It's over there.

70:32

>> Told you there was another panic attack.

70:34

>> Jack,

70:35

>> it's over, man.

70:36

>> They have to get access to the computers

70:39

to replicate themselves. And people are

70:41

still in charge of that. But in

70:43

principle,

70:44

>> once they've got control of the data

70:45

centers, they can replicate themselves

70:47

as much as they like.

70:48

>> Okay. Okay. I got another question. I

70:50

served on a board of the Pentagon for

70:51

like seven years, and it was when AI was

70:55

manifesting itself as a possible tool of

70:59

warfare. And we introduced guidance for

71:05

the invocation of AI in situations that

71:08

the military might encounter. One of

71:09

which was if AI decides

71:13

that it can or should take action that

71:15

will end in death of the enemy, should

71:19

we give it that access to do so

71:21

>> or still a big um debate

71:24

>> or should we always ensure that there's

71:27

a human inside that loop?

71:29

>> It's a big

71:29

>> Okay, so we said there's got to if AI

71:32

cannot make an make its own decision to

71:35

kill, right?

71:36

>> A human has to be in there. My question

71:38

to you is Jeffrey, if there are other

71:40

nations who put in no such safeguards,

71:44

then that is a timing advantage that an

71:47

enemy would have over you.

71:49

>> Correct.

71:50

>> And then we have we have we have one

71:52

more step in the loop that they don't.

71:53

>> Absolutely. But I my belief is that the

71:57

US military isn't committed to the

71:59

always being a human involved in each

72:01

decision to kill. They what they say is

72:03

there will always be human oversight,

72:06

>> right? But in the heat of battle, you've

72:08

got a drone that's going up against a

72:11

Russian tank, and you don't have time

72:14

for a human to say, "Is it okay for the

72:16

drone to drop a grenade on this

72:18

soldier?" So, my suspicion is the US

72:21

military, if you made the

72:23

recommendation, there should always be a

72:24

person.

72:25

>> Well, that was like eight years ago.

72:26

Yeah.

72:27

>> Yeah. I don't think they stand by that

72:28

anymore. I think what they say is

72:30

there'll always be human oversight,

72:32

which is a much vagger thing.

72:34

>> All right. So,

72:34

>> human accountability. On the subject of

72:36

war, is there likely to be international

72:39

cooperation on development of guardrails

72:42

and a human factor in decision-m or is

72:45

this just wild west?

72:47

>> Okay, if you ask when do people

72:49

cooperate, people cooperate when their

72:52

interests are aligned. So at the height

72:54

of the cold war, the USA and the USSR

72:57

cooperated on not having a global

73:00

thermonuclear war because it wasn't in

73:02

either of their interests. Their

73:03

interests were aligned. So if you look

73:05

at the risks of AI, there's using AI to

73:09

corrupt elections with fake videos.

73:12

>> The country's interests are

73:13

anti-aligned. They're all doing it to

73:15

each other,

73:15

>> right?

73:16

>> There's cyber attacks. Their interests

73:19

are basically anti-aligned. There's

73:21

terrorist creating viruses where their

73:23

interests are probably aligned. So they

73:25

might cooperate there. And then there's

73:26

one thing where their interests are

73:28

definitely aligned and they will

73:29

cooperate which is preventing AI from

73:32

taking over from people. If the Chinese

73:35

figured out how you could prevent AI

73:37

from ever wanting to take over, from

73:39

ever wanting to take control away from

73:41

people, they would immediately tell the

73:42

Americans because they don't want AI

73:45

taking control away from people in

73:46

America either. We're all in the same

73:48

boat when it comes to that.

73:50

>> This is the AI version of uh nuclear

73:53

winter.

73:54

>> Yes,

73:55

>> it seems to me

73:56

>> it is. It's exactly that. will cooperate

73:58

to try and avoid that.

74:00

>> Because in nuclear winter, just to

74:01

refresh people's memory, the idea was if

74:03

there's total nuclear exchange, you

74:06

incinerate forests and land and what

74:08

have you. The soot gets into the

74:10

atmosphere, block sunlight, and all life

74:13

dies.

74:14

>> So there is no winner,

74:15

>> of course,

74:16

>> in a total exchange of nuclear weapons.

74:19

>> Mutually assured destruction.

74:21

>> Yeah. And so who wants that?

74:23

>> Unless unless you're a madman or

74:25

something, they exist. Maybe I think

74:27

maybe the cockroaches win.

74:29

>> They win.

74:29

>> Oh, yeah. Well, how about that?

74:31

>> Yeah. This doesn't factor in a possible

74:34

leader who is in a death cult.

74:37

>> A Nero, so to speak.

74:38

>> Yeah. If I moder if I say I don't mind

74:40

if everybody dies cuz I'm going to this

74:42

place when in in death and all my

74:45

followers are coming with me in this

74:46

cult. So that that complicates this

74:49

aligned vision statement that you're

74:52

describing.

74:53

It does complicate it a lot. And I find

74:55

it very comforting that um it's obvious

74:58

that Trump doesn't actually believe in

74:59

God.

75:00

>> Oh,

75:01

let me follow that up with a quote from

75:03

Steven Weinberg.

75:05

>> Okay.

75:05

>> Do you know this quote, Jeffrey?

75:07

>> No.

75:07

>> Steven Weinberg. There will always be

75:09

good people and bad people in the world.

75:12

But to get a good person to do something

75:14

bad requires religion.

75:17

>> That's that's

75:18

>> because they're doing it in the name of

75:19

religion. You did do it in the name of

75:20

some point of anything.

75:22

>> I think we need to we need to recognize

75:24

at this point that we have a religion.

75:26

We call it science. Now it does differ

75:28

from the other religions. And the way it

75:30

differs is it's right.

75:35

>> Mic drop. Okay. Um

75:37

>> wait a minute. I think we got to give

75:39

Jeffrey Hinton like the Turring Prize

75:42

and I give Would you give him a Nobel

75:43

Prize for what he's contributed here?

75:44

>> Well, to go with his other one.

75:47

>> Yes.

75:47

>> No. No. I I I like earrings.

75:50

>> I left that out at the beginning, sir.

75:53

In 2018, you won the Turing prize. This

75:56

is a highly coveted computer science

75:58

prize. Uh, correct. And and and Turing,

76:01

we mentioned him at the beginning of the

76:03

top of the show. So, first

76:04

congratulations on that. And then that

76:06

wasn't enough.

76:08

>> Okay. Uh, the Nobel Committee

76:10

>> sluming with the Nobel.

76:13

>> Yeah. So the Nobel committee said this

76:15

AI stuff that was birthed by by

76:18

Jeffrey's work from decades ago is so

76:21

fundamental to what's going on in this

76:22

world. We got to give this man Nobel

76:24

Prize and earn the Nobel Prize in

76:26

physics 2024.

76:28

>> Just a little correction, there are a

76:30

whole bunch of people birthed AI. Um in

76:32

particular, the back propagation

76:34

algorithm was reinvented by David

76:37

Rumlhart who got a nasty brain disease

76:39

and died young

76:40

>> but he doesn't get enough credit. Oh,

76:43

okay. Thanks for calling that out. Plus,

76:45

the Nobel Committee does not offer a

76:48

Nobel Prize

76:49

>> to you if you're already dead.

76:50

>> So, there's no

76:51

>> You have to be alive when they announce

76:52

it.

76:52

>> Award. No. Well, you can get it if you

76:55

died between when they announced it and

76:56

the ceremony, but not if So, anyway, so

77:00

congratulations on that. And I don't

77:03

mean to brag on our podcast, but you're

77:06

like the fifth Nobel laurate we've

77:08

interviewed.

77:09

>> More than that.

77:09

>> Yeah. Yeah. I think we Yeah. I don't

77:11

mean to brag on our podcast. Yeah,

77:13

that's all.

77:13

>> That's cool, though.

77:14

>> That's cool. Go. Okay.

77:15

>> I have a a follow-up question. I mean,

77:17

we've we've got into the apocalyptic

77:19

scenario and at the moment, hopefully,

77:21

it's a scenario that doesn't play out

77:23

because we are competitive by nature as

77:26

humans and particularly here in the US,

77:28

who is leading the race in artificial

77:32

intelligence and who is likely to cross

77:34

the finish line first when it comes to

77:37

the prize? If I had to bet on one lot of

77:40

people,

77:41

>> Mhm.

77:42

>> it would probably be Germany, Google.

77:45

But I used to work for Google, so don't

77:47

take me too seriously about that. I have

77:49

a vested interest in them winning. Um,

77:51

Anthropic might win, OpenAI might win. I

77:55

think it's less likely that Microsoft

77:57

will win or that Facebook will win.

77:59

>> Well, we know it won't be Facebook.

78:01

Why do you know that?

78:02

>> I mean, let's look at who's running

78:04

Facebook. Okay, come on.

78:06

>> No, it's not who's running it. it who

78:08

has the resources to get the right

78:10

people to do the work.

78:11

>> All right, Jeffrey, the follow up on

78:13

that is whoever crosses the line first,

78:16

what is their prize? What will be the

78:20

reward for them getting there before?

78:22

>> Wait, back up for a sec. Tell me about

78:24

the value of the stock market in the

78:26

last year.

78:28

Okay. And my belief is just from reading

78:31

it in the media that 80% of the increase

78:34

of the value in the stock market, the US

78:36

stock market can be attributed to the

78:39

increase in value of the big AI

78:41

companies.

78:42

>> True.

78:42

>> 80% of the growth.

78:45

>> Yes.

78:45

>> Anyone thinking bubble? And that's kind

78:48

of what they're calling it, the AI

78:50

bubble.

78:52

>> Okay.

78:52

>> The issue is this. There's two senses of

78:55

bubble. One sense of bubble is it turns

78:58

out AI doesn't really work as well as

79:00

people thought it might.

79:02

>> Right?

79:02

>> It doesn't actually develop the ability

79:04

to replace all human intellectual labor

79:07

which is what most people developing it

79:09

believe is going to happen in the end.

79:11

>> That was the fear factor for sure.

79:13

>> Yeah.

79:13

>> The other sense of bubble is the

79:15

companies can't get their money back

79:17

from the investments. Now that seems to

79:20

be more likely kind of bubble

79:22

>> because as far as I understand it, the

79:25

companies are all assuming if we can get

79:27

there first, we can sell people AI that

79:30

will replace a lot of jobs. And of

79:34

course, people will pay a lot of money

79:35

for that. So, we'll get lots of money.

79:38

But they haven't thought about the

79:39

social consequences. If they really do

79:42

replace lots of jobs, the social

79:44

consequences will be terrible.

79:45

>> Correct.

79:46

>> Totally. However, it'll be it'll be

79:48

>> they replace the jobs and now you still

79:50

want to sell your product and no one has

79:52

income to buy the product.

79:54

>> Yeah. It's it's a self-limiting path.

79:57

>> That's the Keynesian view of it. And

79:58

then the additional view is that

80:01

there'll be high unemployment levels

80:03

which will lead to a lot of social

80:04

unrest. So the uh yeah the secondary uh

80:07

view of that is you just have two tiers

80:09

of existence for our societies and the

80:12

first tier is all the people who are

80:14

benefiting from AI and the second tier

80:17

are the you know the the feudal peasants

80:20

that are now forced to live their lives

80:22

because of AI.

80:23

>> Let me ask you a non-AI question because

80:25

just you're a deep thinker in this

80:26

space. That's what everybody said in the

80:29

dawn of automation. Everyone will be

80:31

unemployed. there'll be no jobs left and

80:34

society will go to ruin. Yet society

80:37

expanded with other needs and other

80:40

things people that's why 90% of us are

80:42

no longer farmers. Okay, we we we've

80:45

have machines to do that and we invent

80:47

other things like vacation resource

80:49

>> but that decades this is going to take a

80:51

fraction.

80:52

>> Is that so Jeffrey is the problem here

80:55

the rapidity with which we may create an

80:59

unemployment an unemployed class where

81:01

the society cannot recover from the rate

81:04

at which people are losing their jobs.

81:06

That certainly is one big aspect of the

81:08

problem. But there's another aspect

81:10

which is if you use a tractor to replace

81:13

physical labor, you need far fewer

81:15

people now. Other people can go off and

81:18

do intellectual things. But if you

81:20

replace human intelligence,

81:23

where are they going to go? Where are

81:25

people who work in a call center going

81:27

to go when an AI can do their job

81:30

cheaper and better?

81:32

>> Right. Yeah. This is

81:33

>> Oh, so there's not another thing.

81:35

there's not another thing.

81:36

>> They open another thing and then AI will

81:37

do that.

81:38

>> Right?

81:38

>> Whatever thing you open, AI can do.

81:41

>> You can look at human history in an

81:42

interesting way as getting rid of

81:45

limitations.

81:46

>> So a long time ago, we had the

81:48

limitation you had to worry about where

81:49

your next meal was coming from,

81:51

>> right?

81:51

>> Agriculture got rid of that. It

81:53

introduced a lot of other problems, but

81:54

it got rid of that particular worry.

81:56

Then we had the limitation you couldn't

81:58

travel very far. Well, the bicycle

82:00

helped a lot with that and cars and

82:03

airplanes. We got over that kind of

82:05

limitation. For a long time, we had the

82:07

limitation. We were the ones who had to

82:09

do the thinking. We're just about to get

82:11

over that limitation.

82:13

And it's not clear what happens once you

82:15

got over all the limitations. People

82:17

like Sam Elman think it'll be wonderful,

82:19

>> right? So, we we'll become AI's pet.

82:23

>> Well, no. A lot of people believe that

82:25

this is the um and this this movement

82:27

started years ago for universal global

82:30

income.

82:31

>> Okay. So would you say Jeffrey that the

82:33

the universal basic income the stock

82:35

value the figurative stock value in that

82:38

idea is growing as AI gains power.

82:42

>> It's becoming to seem more essential but

82:44

it has lots of problems. So one problem

82:47

is many people get their sense of

82:48

selfworth from the job they do and it

82:51

won't deal with the dignity issue.

82:53

Another problem is the tax base. If you

82:55

replace workers with AIs, the government

82:58

loses its tax base. It has to somehow be

83:02

able to tax the AIs. But the big

83:04

companies aren't going to like that.

83:06

>> I think we should let AI figure out this

83:07

problem.

83:08

>> That's right.

83:13

So Jeffrey the many people uh especially

83:16

sci-fi writers distinguish between the

83:19

power and intellect of machines fine and

83:23

the crossover when they become conscious

83:27

and that's was a big moment in the

83:31

Terminator series

83:32

>> that was the singularity in the

83:33

terminator

83:34

>> when Skynet Skynet

83:35

>> had enough neural connections or

83:37

whatever kind of connections made it so

83:40

that it achieved consciousness. So there

83:42

seems to be and if you come to this as a

83:44

as a cognitive psychologist, I'm curious

83:46

how you think about this. Are we allowed

83:48

to presume that given sufficient

83:51

complexity in any neural net be it real

83:54

or imag or or artificial something such

83:57

as consciousness emerges.

83:59

>> So the problem here is not really a

84:02

scientific problem. It's that most

84:05

people in our culture have a theory of

84:08

how the mind works and they have a view

84:10

of consciousness as some kind of essence

84:12

that emerges. I think consciousness is

84:15

like flegiston maybe. Um it's an essence

84:18

that's designed to explain things and

84:22

once we understand those things we won't

84:23

be trying to use that essence to explain

84:25

them. I want to try and convince you

84:27

that a multimodal chatbot already has

84:30

subjective experience. So people use the

84:32

word sentience or consciousness or

84:34

subjective experience. Let's focus on

84:37

subjective experience for now. Most

84:39

people in our culture think that the way

84:42

the mind works is it's a kind of

84:43

internal theater. And when you're doing

84:45

perception, the world shows up in this

84:48

internal theater and only you can see

84:50

what's there. So if I say to you, if I

84:53

drink a lot and I say to you, I have the

84:55

subjective experience of little pink

84:57

elephants floating in front of me. Most

84:59

people interpret that as there's this

85:01

inner theater, my mind and I can see

85:04

what's in it and what's in it is little

85:05

pink elephants and they're not made of

85:08

real pink and real elephants. So they

85:10

must be made of something else. So

85:11

philosophers invent qualia which is kind

85:13

of the flegiston of cognitive science.

85:16

They say they must be made of qualia.

85:18

Let me give you a completely different

85:20

view that is Daniel Dennett's view who

85:22

was a great philosopher of cognitive

85:23

science which is

85:24

>> late great philosopher. Yeah,

85:25

>> the late great that view of the mind is

85:28

just utterly wrong. So I'm now going to

85:31

say the same thing as when I told you I

85:34

had the subjective experience of Olympic

85:35

elephants without using the word

85:37

subjective experience and without

85:39

appealing to Qualia. I start off by

85:42

saying I believe my perceptual systems

85:45

lying to me. That's the subjective bit

85:48

of it. But if my perceptual system

85:50

wasn't lying to me, there would be

85:52

little pink elephants out there in the

85:54

world floating in front of me. So what's

85:56

funny about these little pink elephants

85:57

is not that they're made of qualia and

85:58

they're in an inner theta. It's that

86:00

they're hypothetical. They're a

86:02

technique for me telling you how my

86:05

perceptual systems lying by telling you

86:07

what would have to be there for my

86:09

perceptual system to be telling the

86:11

truth. And now I'm going to do it with a

86:13

chatbot. I take a multimodal chatbot. I

86:16

train it up. It's got a camera. It's got

86:17

a robot arm. It can talk. I put an

86:20

object in front of it and I say, "Point

86:21

at the object and it points at the

86:23

object." Then I mess up its perceptual

86:25

system. I put a prism in front of the

86:28

camera. And now I put an object in front

86:29

of it and say, "Point at the object."

86:31

And it points off to one side. And I say

86:33

to it, "No, that's not where the object

86:35

is. It's actually straight in front of

86:36

you." But I put a prism in front of your

86:38

lens. And the chatbot says, "Oh, I see.

86:41

The prism bent the light rays, so the

86:43

object is actually straight in front of

86:45

me." But I had the subjective experience

86:47

that it was off to one side. Now, if the

86:49

chatbot said that, it would be using

86:51

words subjective experience exactly the

86:53

way we use them. And so that chatbot

86:56

would have just had a subjective

86:57

experience.

86:59

>> Now, what if you um first went out

87:01

drinking with the chatbot and you had a

87:04

very significant amount of Johnny Walker

87:06

Blue?

87:08

>> That's extremely improbable. I would

87:10

have Leafrog.

87:11

>> Oh. Oh.

87:13

>> Oh, you're I see you're an eye man. You

87:15

like the piness of the leaf. Okay, good

87:18

man.

87:18

>> Oh, so if I understand what you just

87:22

shared with us in these two examples,

87:25

>> you actually pulled a consciousness

87:27

touring test on us. You said a human

87:30

would do this and now your chatbot does

87:33

it and it's fundamentally the same. So

87:37

if you want to say we're conscious for

87:39

exhibiting that behavior, you're going

87:41

to have to say the chatbot's conscious

87:43

and inventing whatever mysterious fluid

87:46

is making that happen. But it could be

87:48

that we are the whole concept of

87:51

consciousness is a distraction from just

87:53

the actions that people take in the face

87:55

of stimulus.

87:56

>> Okay. So notice that the chatbot doesn't

87:59

have any mysterious essence or fluid

88:02

called consciousness, but it has a

88:03

subjective experience just like we do.

88:05

So I think this whole idea of

88:07

consciousness is some magic essence that

88:10

you suddenly get indicted with if you're

88:11

complicated enough is just nonsense.

88:14

>> Yeah, there you go.

88:16

>> I agree. I've always felt that

88:18

consciousness was something people are

88:19

trying to explain without knowing if it

88:21

really exists

88:23

>> in in any kind of tangible way,

88:24

>> which is why it's always difficult to

88:25

describe because you don't know what it

88:26

is

88:27

>> for example. Yes. Yes. But I think there

88:29

is awareness. And if you look at what

88:32

scientists say when they're not thinking

88:34

philosophically,

88:35

there's a lovely paper where the chatbot

88:38

says, "Now, let's be honest with each

88:40

other. Are you actually testing me?" And

88:42

the scientists say, "The chatbot was

88:44

aware it was being tested." So, they're

88:47

attributing awareness to a chatbot. And

88:49

in everyday conversation, you call that

88:51

consciousness. It's only when you start

88:53

thinking philosophically and thinking

88:54

that it's some funny mysterious essence

88:57

that you get all confused.

88:58

>> Well, there is

88:59

>> I have to say that this has been a

89:01

fascinating conversation that will cause

89:03

me not to sleep for a month.

89:05

>> Um, yeah,

89:06

>> you get plenty of work done.

89:09

>> So, Jeffrey, take us out on a positive

89:12

note, please. So, we still have time to

89:15

figure out if there's a way we can

89:17

coexist happily with AI and we should be

89:20

putting a lot of research effort into

89:22

that because if we can coexist happily

89:24

with it and we can solve all the social

89:27

problems that will arise when it makes

89:29

all our jobs much easier then it can be

89:31

a wonderful thing for people.

89:33

>> Agreed. Okay. So, so there is hope.

89:36

>> Yes. And one last thing because you

89:38

hinted at it, this point of singularity

89:41

where AI trains on itself

89:45

so that it exponentially gets smarter

89:47

like by the minute. That's been called a

89:50

singularity by many people. Of course,

89:52

Ray Kershw among them who's been a guest

89:55

on a previous episode of Stars.

89:56

>> Yeah. A couple of times. Yeah. So, what

89:59

is your sense of this singularity? Is it

90:01

real the way others say? Is it imminent

90:04

the way others say?

90:06

I don't know the answer to either of

90:08

those questions. My suspicion is AI will

90:10

get better at us in the end at

90:13

everything, better than us at

90:14

everything, but it'll be sort of one

90:16

thing at a time. It's currently much

90:18

better than us at chess and go. It's

90:20

much better than us at knowing a lot of

90:22

things. Not quite as good as us at

90:24

reasoning. I think rather than sort of

90:26

massively overtaking us in everything

90:28

all at once, it'll be done one area at a

90:31

time. And my sort of way out of that is,

90:35

you know, I get to walk a beach and look

90:37

at pebbles and seashells. AI doesn't.

90:40

>> Yeah. It can create its own beach.

90:43

>> No. Would it only know about the new

90:45

mollisk that I discovered if I write it

90:48

up and put it online?

90:49

>> Mhm.

90:50

>> So, the human can continue to explore

90:52

the universe in ways that AI doesn't

90:56

have access to.

90:57

>> There's one word missing from your

90:58

entire assessment.

90:59

>> What's that?

91:00

>> Yet.

91:05

Yeah, I just think of my, you know, will

91:07

AI come up with a new theory of the

91:09

universe that requires human insights

91:12

that it doesn't have because I'm

91:14

thinking the way no one has thought

91:16

before.

91:17

>> I think it will.

91:19

>> That's not the answer I wanted from you.

91:20

>> Yeah, I was.

91:22

>> But that's the answer you got.

91:23

>> Let me give you an example. AI is very

91:25

good at analogies already. So when chat

91:28

GPD4 was not allowed to look on the web

91:31

when all its knowledge was in its

91:32

weights, I asked it why is a compost

91:34

heap like an atom bomb and it knew it

91:38

said the energy scales are very

91:39

different and the time scales are very

91:41

different. But it then went on to talk

91:43

about how when a compost heap gets

91:44

hotter it generates heat faster and when

91:46

an atom bomb generates more neutrons it

91:48

generates neutrons faster. Um, so it

91:51

understood the commonality and it had to

91:53

understand that to pack all that

91:55

knowledge into so few connections, only

91:57

a trillion or so. That's a source of

91:59

much creativity

92:00

>> and it's not just by finding words that

92:02

were juxtaposed with other words.

92:04

>> No, it understood what a chain reaction

92:06

was.

92:07

>> Yeah.

92:08

>> Well, all right. That's the end of us.

92:11

>> Yeah. We're done

92:12

>> on Earth. We're done.

92:13

>> We're finished.

92:14

>> This is the last episode. We

92:16

>> stick in us. We're done. Gentlemen, it's

92:19

been a pleasure.

92:23

>> Well, Jeffrey Hinton, it's been a

92:24

delight to have you on.

92:26

>> We know you're you're tugged in many

92:28

directions, especially after your recent

92:30

Nobel Prize, and we're delighted you

92:33

gave us a piece of your surely

92:35

overscheduled and busy life.

92:37

>> Thank you for inviting me.

92:38

>> Well, guys, that was something.

92:41

>> Did you sit comfortably through all of

92:42

that?

92:43

>> I was I I I squirmed. I squirmed.

92:46

>> I knew you'd panic. Well, no. I have to

92:47

tell you that um certain parts of the um

92:50

conversation gave me the anxiety of, you

92:52

know, sitting in a theater theater with

92:54

diarrhea.

92:56

>> Thanks for that explicit.

92:58

>> Thanks for sharing. That That's the

93:00

nicest thing anybody's ever said about

93:02

me.

93:07

>> On that note, this has been Star Talk

93:10

special edition. Chuck, always good to

93:12

have you. Gary, love having you right at

93:15

my side. Neil deGrasse Tyson bidding you

93:17

as always to keep looking up however

93:21

much harder that will become.

UNLOCK MORE

Sign up free to access premium features

INTERACTIVE VIEWER

Watch the video with synced subtitles, adjustable overlay, and full playback control.

SIGN UP FREE TO UNLOCK

AI SUMMARY

Get an instant AI-generated summary of the video content, key points, and takeaways.

SIGN UP FREE TO UNLOCK

TRANSLATE

Translate the transcript to 100+ languages with one click. Download in any format.

SIGN UP FREE TO UNLOCK

MIND MAP

Visualize the transcript as an interactive mind map. Understand structure at a glance.

SIGN UP FREE TO UNLOCK

CHAT WITH TRANSCRIPT

Ask questions about the video content. Get answers powered by AI directly from the transcript.

SIGN UP FREE TO UNLOCK

GET MORE FROM YOUR TRANSCRIPTS

Sign up for free and unlock interactive viewer, AI summaries, translations, mind maps, and more. No credit card required.