TRANSCRIPTEnglish

AI, Gods, and Selves: Incredibly Effective Illusions

55m 17s6,591 words1,084 segmentsEnglish

FULL TRANSCRIPT

0:01

The rise of AI makes people worry. It

0:04

causes moral concern. We fear a loss of

0:07

agency, a kind of existential

0:10

humiliation.

0:12

On the one hand, we treat AI as if it

0:15

was an intelligent agent who understands

0:18

and guides us. When talking to an LLM

0:21

like ChatGpt or DeepSeek, it feels like

0:25

we're talking to someone. But on the

0:27

other hand, we know this someone isn't

0:30

like us. That's quite uncanny.

0:34

AI challenges our sense of self and our

0:39

existential orientation.

0:42

In this video, I'll trace the history of

0:44

our sense of self and of our sense of AI

0:49

to point out that AI, the self and the

0:53

gods, for that matter, are incredibly

0:56

effective illusions which don't really

0:59

exist.

1:01

I will also argue at the end of the

1:03

video against dystopian visions of a

1:06

singularity

1:08

and question some current approaches to

1:11

AI ethics.

1:13

The video is inspired by a recent

1:16

conversation with Iad Rafa, director of

1:19

the center for humans and machines at

1:21

the Maxplank Institute for Human

1:24

Development in Berlin. And it's also

1:27

inspired by the many conversations with

1:29

Milo Jeanich.

1:32

Why do we speak of artificial

1:34

intelligence although we know very well

1:37

it's not really intelligent in the sense

1:40

of a thinking thing. It even tells us

1:43

so. Deepseek says do I have

1:46

consciousness, emotions or

1:48

self-awareness? No. And it adds that it

1:52

doesn't think that this is just a figure

1:54

of speech.

1:56

When I use words like understanding,

1:59

thinking or knowing, we are using them

2:01

as metaphors or analogies.

2:05

Interesting by the way that it says I

2:08

here first and then we. And in fact,

2:12

it's I is actually nothing but our we.

2:16

But I will get to this issue later.

2:19

Of course, we know that it is not anyone

2:22

and doesn't think. But we still feel

2:26

understood by it. Why? When prompted,

2:30

and I edited out here all the flattering

2:32

language meant to seduce us and keep us

2:34

on the platform, Deep Seek says what we

2:38

already know. For millennia, the only

2:42

thing capable of producing meaningful

2:45

language was a conscious human mind. So

2:48

we are deeply conditioned to equate

2:51

fluent language use with an inner

2:54

conscious life.

2:57

When further prompted, deep admits that

3:01

ascribing intelligence to it in the

3:03

sense of treating it as if it were

3:05

really a mind is an incredibly effective

3:10

illusion.

3:11

Let's take note of this. We stick to the

3:15

metaphor of artificial intelligence

3:18

because we are conditioned to the

3:21

effective illusion that when we

3:23

communicate with something, it must be

3:26

an intelligent agent.

3:30

Apparently, in order to be socially and

3:32

psychologically effective, intelligence

3:35

must be ascribed to a being that is

3:38

intelligent.

3:39

And this intelligent being can talk.

3:42

It's someone who communicates, someone

3:45

whom we listen to and who makes sense.

3:49

But who precisely this envisioned

3:52

thinking and speaking thing is is

3:56

historically in the long run contingent

4:00

to trace the history of this envisioned

4:02

thing. This video looks into some

4:05

contemporary theory. Elena Esposto,

4:09

Nicholas Lumman, Julian James, and Uval

4:12

Harrari. A while ago on this channel, we

4:15

posted a conversation with Elena Posto

4:18

about her book artificial communication.

4:22

The MIT press website where you can

4:24

download a book for free says in

4:27

artificial communication Elena Esposito

4:30

argues that the analogy between

4:33

algorithms and human intelligence is

4:36

misleading. Esposito proposes that we

4:40

think of smart machines not in terms of

4:43

artificial intelligence but in terms of

4:46

artificial communication. But her

4:49

proposition did not catch on. No one

4:52

really says artificial communication as

4:55

positive suggestion to understand AI in

4:58

social rather than in mental terms is

5:01

informed by her academic teacher

5:04

Nicholas Lumman. I consider Lumman's

5:07

systems theory the most advanced theory

5:11

framework available today. It's the only

5:14

theory I know perhaps next to NZA but he

5:18

doesn't really have a theory that

5:21

decisively beyond what Lumman calls old

5:24

European thought that is beyond

5:28

enlightenment and its humanism.

5:31

Here are some basics of the theory that

5:33

I hope you allow me to introduce before

5:36

presenting my main point. At the heart

5:39

of it is a distinction between three

5:41

kinds of autopoadic self-reproducing

5:45

systems. Three kinds of organisms so to

5:48

say. There are biological systems. The

5:52

self-reproductive evolution of life like

5:55

our bodies or plants. They consist of

5:59

all kinds of physiological processes.

6:02

They are mental systems. The self

6:05

reproductive evolution of consciousness.

6:08

what we call minds. They consist of all

6:12

kinds of psychological processes, their

6:15

social systems, the self reproductive

6:18

evolution of communication like the

6:21

economic system, the political system or

6:24

the media system. They consist of all

6:27

kinds of social processes like payments,

6:31

elections or YouTube videos.

6:34

All these systems are environments for

6:37

one another. To record this video, to be

6:41

able to talk to you on social media, my

6:44

body must be alive and my mind needs to

6:48

think. And for you, the same is the case

6:51

if you're watching it. Communication,

6:54

including social media, operates within

6:58

the environment of bodies and minds.

7:01

The three systems co-evolve.

7:05

They're often structurally coupled. As

7:07

Lumman says, society influences how

7:11

minds evolve and minds influence how

7:15

society evolves. And this in turn also

7:18

influences biological evolution in

7:20

various ways. That's by the way why we

7:23

speak of the entroposy.

7:26

In other words, systems are contingent

7:29

upon one another.

7:31

Curiously and importantly, systems are

7:35

operationally closed. They're not in

7:38

direct mechanical contact with one

7:40

another. They live on their own, so to

7:42

say. This video is social media

7:46

communication. I cannot continue

7:49

recording it and you cannot continue

7:52

watching it simply by being bodily alive

7:56

and mentally thinking. We need to

8:00

communicate to keep communication going.

8:04

I need to post the video and you need to

8:07

watch it. Communication can only be

8:10

continued by more communication, not by

8:14

biological or mental operations on their

8:18

own.

8:20

Systems operate only with their own

8:23

operations, not with the operations of

8:26

other systems. However, systems irritate

8:30

one another. What I say irritates what

8:34

you think. It triggers thoughts which in

8:37

turn can irritate your body. When I

8:40

speak, you may be scratching your head

8:42

or start yawning or maybe even smile

8:45

slightly.

8:47

Minds in society are coupled by the

8:50

common medium of language. Mental

8:54

systems are intelligent, can think in

8:56

language.

8:58

Social systems can communicate in the

9:01

same language and yet as mentioned they

9:05

remain operationally close to one

9:07

another. Thoughts never become part of

9:11

communication as thoughts. They must be

9:15

communicated and communication never

9:19

becomes part of thought as

9:21

communication. You must think on it.

9:25

Ivan not translator of Lumman's book

9:29

social systems calls this systemic rift

9:32

between minds and communication

9:35

hermeneutic despair and she illustrates

9:38

it with a scene from a theater play. She

9:42

writes

9:43

in Danton's death bushna dramatizes the

9:48

primal scene of hermeneutic despair. The

9:51

protagonist makes a silent gesture

9:54

toward his lover's forehead and says to

9:57

her, "There, there, what lies behind

9:59

this?" To understand one another, we

10:03

would have to break open each other's

10:06

skulls and pull the thoughts out of the

10:09

fibers of our brains.

10:13

Lumman writes, "Categorically,

10:16

human beings cannot communicate. Not

UNLOCK MORE

Sign up free to access premium features

INTERACTIVE VIEWER

Watch the video with synced subtitles, adjustable overlay, and full playback control.

SIGN UP FREE TO UNLOCK

AI SUMMARY

Get an instant AI-generated summary of the video content, key points, and takeaways.

SIGN UP FREE TO UNLOCK

TRANSLATE

Translate the transcript to 100+ languages with one click. Download in any format.

SIGN UP FREE TO UNLOCK

MIND MAP

Visualize the transcript as an interactive mind map. Understand structure at a glance.

SIGN UP FREE TO UNLOCK

CHAT WITH TRANSCRIPT

Ask questions about the video content. Get answers powered by AI directly from the transcript.

SIGN UP FREE TO UNLOCK

GET MORE FROM YOUR TRANSCRIPTS

Sign up for free and unlock interactive viewer, AI summaries, translations, mind maps, and more. No credit card required.

    AI, Gods, and Selves:… - Full Transcript | YouTubeTranscript.dev