文本记录English

MIT 6.S087: Foundation Models & Generative AI. CHAT-GPT & LLMs

1h 5m 6s11,635 字数1,658 segmentsEnglish

完整文本记录

0:00

all

0:01

right welcome to the third lecture on

0:06

Foundation mulative AI So today we're

0:09

going to cover chat

0:11

GPT um and um right I mean I think for a

0:16

lot of people chat GP was

0:18

the the tool or the the AI that really

0:23

made people understand this is different

0:25

now we're able to do things we weren't

0:27

able to do before and and definitely uh

0:30

created some kind of hype uh so

0:33

hopefully after this lecture you'll you

0:35

understand kind of the basic idea and

0:37

also somehow understand the BET right

0:39

the bet that open Ai and Ilia the head

0:42

researcher did in terms of what actually

0:44

would lead to CHP and how in hindsight

0:47

it might be quite I mean easy but it was

0:52

a really daring bad not obvious at all

0:55

at the time that this would actually

0:56

work out

0:58

um so should be be a lot of fun and just

1:01

to quickly go through our course

1:02

schedule as well a little bit right so

1:04

today is January 16 uh and next time

1:07

we'll talk about stable diffusion image

1:10

generation and then we'll talk about

1:12

emerging Foundation models basically

1:14

Foundation models generative AI in the

1:16

commercial space H we'll have two guest

1:19

speakers and then we'll end with the

1:21

lecture on AI ethics and regulation as

1:24

well as a

1:26

panel okay so what have we talked about

1:28

before we started off

1:30

H with an introduction a short high

1:33

level intuitive answer to what is

1:35

foundation M generative

1:36

AI we went a little bit on a

1:38

philosophical digression and asked about

1:41

how's the world structured because that

1:43

allows us to think about how we should

1:44

learn in the world then we on the second

1:47

lecture went through all the different

1:48

algorithms um and yeah today we'll we'll

1:52

dive in more specifically into chpt and

1:54

kind of uh pull everything together um

1:58

and to reiterate right so what do we do

2:01

in uh Foundation models geni well we

2:04

apply this self-supervised learning

2:06

where we learn without uh label data so

2:09

we can we can get you know as much data

2:12

as we want because there's no human

2:13

being in the loop so there's no limit

2:15

how much we can scale this up and and

2:18

what we get from this you know by

2:19

learning from observation and learning

2:21

from the data directly is a very

2:23

contextual and relational understanding

2:25

of meaning and we gave this example

2:27

before about you know from a supervised

2:30

learning perspective you learn what a

2:32

dog is from seeing you know labeled uh

2:36

examples of dogs and in reinforcement

2:39

learning you focus on optimizing certain

2:41

goals and you understand a dog in

2:43

relation to how it makes you happy or

2:45

fulfilled in some sense or optimizing

2:47

your goals but in self supervised

2:49

learning right it's the foundational

2:50

technology behind uh Foundation models

2:53

you learn from observing dogs in

2:54

different context and you get a very

2:56

relational definition of a dog so it's

2:59

something that's walk by an owner with a

3:00

leash it has an anistic R with cats it

3:03

chases fris with oone right this is your

3:06

definition of what a dog is and today

3:09

we'll you talk about something that's

3:11

extremely engineering heavy in you know

3:15

chat GPT uh relies on a lot of tricks

3:19

and Engineering insights and

3:21

breakthroughs that we're not going to

3:22

cover and I think still though you know

3:26

like it's like talking about a car you

3:28

can understand the high level

3:29

perspective of a car and get some

3:31

insights how to work how it works and

3:32

how it's going to be useful for you

3:34

without getting into all the engineering

3:36

details but of course in real life those

3:39

engineering details really really

3:40

matters and are very very hard to get

3:41

right and that's something that we won't

3:43

really dive into in this lecture because

3:46

that's just when you bring something up

3:48

certain scale and you have to paralyze a

3:49

lot of machines Etc and think about high

3:52

parameters it's a whole science so it's

3:54

not trival at all but it's kind of hard

3:57

uh to teach in a course like this and

3:59

and you have to learn by just actually

4:01

building this

4:03

stuff

4:05

um okay so

4:09

um also a little bit of philosophizing

4:12

in this uh class as

4:15

well

4:17

um I think that again like we talked

4:21

about a little bit of a theme here right

4:22

is that the why this new AI is so

4:25

powerful is because it doesn't Force

4:28

things to comply to Simple Rules right

4:32

it kind of abandons our ability to

4:34

understand and compress what we're

4:35

seeing and deals with that chaos

4:37

directly that's why AI is so powerful

4:38

and so

4:39

humanlike um so also like when I talk

4:42

about this in CHP we try to make very

4:45

high level um statement but of course

4:49

the nuances matters and I think it's

4:51

quite interesting uh I took this quote

4:53

from a general from the 18 and

4:58

1700s and he says this uh quote that P

5:02

Theory which sets itself in opposition

5:04

to the mind and what he meant was that

5:06

he's a general so he fights in battles

5:09

and War and at the time people loved to

5:13

come up and theorize around War like we

5:15

should have certain rules and how

5:16

soldiers should behave in fighting and

5:18

stuff like that but he's like well I've

5:20

been in War uh and Wars don't comply to

5:24

rules first off so you know everybody

5:28

has a plan before they get hit in face

5:29

basically so you know as people start

5:31

shooting at you and you have this fog of

5:33

War of you don't know what's going on

5:35

there's no simple rules to help you

5:37

there and also what he says this in

5:39

terms of the mind he says like well

5:41

actually he's realized by working with

5:43

soldiers that soldiers and human beings

5:45

our mind we're not good at acting

5:49

according to rules that we try to

5:50

memorize we're very intuitive and very

5:52

kind of quick to react to things by our

5:55

intuition that's what really really

5:56

matters and that's what we're strong at

5:57

so if you force a soldier's well Al try

5:59

to memorize a lot of rules and that's

6:01

how it should act in a battle you're

6:03

kind of screwed and very limited in what

6:04

you can do uh which also is something

6:07

that I think AI uh in a new type of AI

6:12

leverages okay so chat

6:16

GPT um right this is a really amazing

6:20

breakthrough that uh has some very

6:22

humanlike Mastery of language that we

6:24

can communicate that can basically solve

6:27

a really wide array of tasks for us

6:30

anything that can be phrased in terms of

6:32

text language it can it can basically

6:35

solve and now as well when with gp4 ET

6:38

becomes uh it's able to handle multi

6:40

modalities but it's it's extremely

6:42

powerful so let's try to break this

6:45

apart well first off what does this name

6:48

actually stand for well the chat part is

6:51

obvious it stands for chat and then GPT

6:53

stands for generative pre-trained

6:56

Transformer and this is a I mean a good

6:59

description of what this uh actually is

7:02

um and I think also if you look at the

7:06

the two different three different

7:08

concepts here they're also almost

7:10

corresponding length in terms of how

7:12

important and influential they are in

7:14

making chat GPT work so chat part we

7:18

we'll cover last it's the kind of the

7:20

least important one in some sense H the

7:22

Genty pre-trained is the self supervised

7:25

step of how you train this and arrive at

7:26

this uh model and then the Transformer

7:29

is the basically the engine behind it in

7:31

some sense

7:33

and so let's start with this generative

7:37

pre-train what does it mean how do we

7:39

pre-train this model and that's

7:41

basically where openi spent 99% of the

7:45

compute was to do this pre-training step

7:47

so it's it's very very

7:50

important okay so what we're going to do

7:52

is that we're going to uh just take some

7:56

random text from the internet so we have

7:57

a sequence of words and and then we're

7:59

just going to try to predict uh the next

8:02

word based on previous words so let's

8:03

say we have uh we start with i here as

解锁更多

免费注册以访问高级功能

互动查看器

观看带有同步字幕、可调节叠加层和完整播放控制的视频。

免费注册以解锁

AI 摘要

获取由 AI 立即生成的视频内容摘要、要点和结论。

免费注册以解锁

翻译

一键将字幕翻译成 100 多种语言。以任何格式下载。

免费注册以解锁

思维导图

将字幕可视化为交互式思维导图。一目了然地了解结构。

免费注册以解锁

与字幕聊天

提出关于视频内容的问题。直接从字幕中获取由 AI 驱动的答案。

免费注册以解锁

从您的字幕中获得更多

免费注册并解锁交互式查看器、AI 摘要、翻译、思维导图等。无需信用卡。

    MIT 6.S087: Foundation Models &… - 完整文字记录 | YouTubeTranscript.dev