TRANSCRIPCIÓNEnglish

MIT 6.S087: Foundation Models & Generative AI. CHAT-GPT & LLMs

1h 5m 6s11,635 palabras1,658 segmentsEnglish

TRANSCRIPCIÓN COMPLETA

0:00

all

0:01

right welcome to the third lecture on

0:06

Foundation mulative AI So today we're

0:09

going to cover chat

0:11

GPT um and um right I mean I think for a

0:16

lot of people chat GP was

0:18

the the tool or the the AI that really

0:23

made people understand this is different

0:25

now we're able to do things we weren't

0:27

able to do before and and definitely uh

0:30

created some kind of hype uh so

0:33

hopefully after this lecture you'll you

0:35

understand kind of the basic idea and

0:37

also somehow understand the BET right

0:39

the bet that open Ai and Ilia the head

0:42

researcher did in terms of what actually

0:44

would lead to CHP and how in hindsight

0:47

it might be quite I mean easy but it was

0:52

a really daring bad not obvious at all

0:55

at the time that this would actually

0:56

work out

0:58

um so should be be a lot of fun and just

1:01

to quickly go through our course

1:02

schedule as well a little bit right so

1:04

today is January 16 uh and next time

1:07

we'll talk about stable diffusion image

1:10

generation and then we'll talk about

1:12

emerging Foundation models basically

1:14

Foundation models generative AI in the

1:16

commercial space H we'll have two guest

1:19

speakers and then we'll end with the

1:21

lecture on AI ethics and regulation as

1:24

well as a

1:26

panel okay so what have we talked about

1:28

before we started off

1:30

H with an introduction a short high

1:33

level intuitive answer to what is

1:35

foundation M generative

1:36

AI we went a little bit on a

1:38

philosophical digression and asked about

1:41

how's the world structured because that

1:43

allows us to think about how we should

1:44

learn in the world then we on the second

1:47

lecture went through all the different

1:48

algorithms um and yeah today we'll we'll

1:52

dive in more specifically into chpt and

1:54

kind of uh pull everything together um

1:58

and to reiterate right so what do we do

2:01

in uh Foundation models geni well we

2:04

apply this self-supervised learning

2:06

where we learn without uh label data so

2:09

we can we can get you know as much data

2:12

as we want because there's no human

2:13

being in the loop so there's no limit

2:15

how much we can scale this up and and

2:18

what we get from this you know by

2:19

learning from observation and learning

2:21

from the data directly is a very

2:23

contextual and relational understanding

2:25

of meaning and we gave this example

2:27

before about you know from a supervised

2:30

learning perspective you learn what a

2:32

dog is from seeing you know labeled uh

2:36

examples of dogs and in reinforcement

2:39

learning you focus on optimizing certain

2:41

goals and you understand a dog in

2:43

relation to how it makes you happy or

2:45

fulfilled in some sense or optimizing

2:47

your goals but in self supervised

2:49

learning right it's the foundational

2:50

technology behind uh Foundation models

2:53

you learn from observing dogs in

2:54

different context and you get a very

2:56

relational definition of a dog so it's

2:59

something that's walk by an owner with a

3:00

leash it has an anistic R with cats it

3:03

chases fris with oone right this is your

3:06

definition of what a dog is and today

3:09

we'll you talk about something that's

3:11

extremely engineering heavy in you know

3:15

chat GPT uh relies on a lot of tricks

3:19

and Engineering insights and

3:21

breakthroughs that we're not going to

3:22

cover and I think still though you know

3:26

like it's like talking about a car you

3:28

can understand the high level

3:29

perspective of a car and get some

3:31

insights how to work how it works and

3:32

how it's going to be useful for you

3:34

without getting into all the engineering

3:36

details but of course in real life those

3:39

engineering details really really

3:40

matters and are very very hard to get

3:41

right and that's something that we won't

3:43

really dive into in this lecture because

3:46

that's just when you bring something up

3:48

certain scale and you have to paralyze a

3:49

lot of machines Etc and think about high

3:52

parameters it's a whole science so it's

3:54

not trival at all but it's kind of hard

3:57

uh to teach in a course like this and

3:59

and you have to learn by just actually

4:01

building this

4:03

stuff

4:05

um okay so

4:09

um also a little bit of philosophizing

4:12

in this uh class as

4:15

well

4:17

um I think that again like we talked

4:21

about a little bit of a theme here right

4:22

is that the why this new AI is so

4:25

powerful is because it doesn't Force

4:28

things to comply to Simple Rules right

4:32

it kind of abandons our ability to

4:34

understand and compress what we're

4:35

seeing and deals with that chaos

4:37

directly that's why AI is so powerful

4:38

and so

4:39

humanlike um so also like when I talk

4:42

about this in CHP we try to make very

4:45

high level um statement but of course

4:49

the nuances matters and I think it's

4:51

quite interesting uh I took this quote

4:53

from a general from the 18 and

4:58

1700s and he says this uh quote that P

5:02

Theory which sets itself in opposition

5:04

to the mind and what he meant was that

5:06

he's a general so he fights in battles

5:09

and War and at the time people loved to

5:13

come up and theorize around War like we

5:15

should have certain rules and how

5:16

soldiers should behave in fighting and

5:18

stuff like that but he's like well I've

5:20

been in War uh and Wars don't comply to

5:24

rules first off so you know everybody

5:28

has a plan before they get hit in face

5:29

basically so you know as people start

5:31

shooting at you and you have this fog of

5:33

War of you don't know what's going on

5:35

there's no simple rules to help you

5:37

there and also what he says this in

5:39

terms of the mind he says like well

5:41

actually he's realized by working with

5:43

soldiers that soldiers and human beings

5:45

our mind we're not good at acting

5:49

according to rules that we try to

5:50

memorize we're very intuitive and very

5:52

kind of quick to react to things by our

5:55

intuition that's what really really

5:56

matters and that's what we're strong at

5:57

so if you force a soldier's well Al try

5:59

to memorize a lot of rules and that's

6:01

how it should act in a battle you're

6:03

kind of screwed and very limited in what

6:04

you can do uh which also is something

6:07

that I think AI uh in a new type of AI

6:12

leverages okay so chat

6:16

GPT um right this is a really amazing

6:20

breakthrough that uh has some very

6:22

humanlike Mastery of language that we

6:24

can communicate that can basically solve

6:27

a really wide array of tasks for us

6:30

anything that can be phrased in terms of

6:32

text language it can it can basically

6:35

solve and now as well when with gp4 ET

6:38

becomes uh it's able to handle multi

6:40

modalities but it's it's extremely

6:42

powerful so let's try to break this

6:45

apart well first off what does this name

6:48

actually stand for well the chat part is

6:51

obvious it stands for chat and then GPT

6:53

stands for generative pre-trained

6:56

Transformer and this is a I mean a good

6:59

description of what this uh actually is

7:02

um and I think also if you look at the

7:06

the two different three different

7:08

concepts here they're also almost

7:10

corresponding length in terms of how

7:12

important and influential they are in

7:14

making chat GPT work so chat part we

7:18

we'll cover last it's the kind of the

7:20

least important one in some sense H the

7:22

Genty pre-trained is the self supervised

7:25

step of how you train this and arrive at

7:26

this uh model and then the Transformer

7:29

is the basically the engine behind it in

7:31

some sense

7:33

and so let's start with this generative

7:37

pre-train what does it mean how do we

7:39

pre-train this model and that's

7:41

basically where openi spent 99% of the

7:45

compute was to do this pre-training step

7:47

so it's it's very very

7:50

important okay so what we're going to do

7:52

is that we're going to uh just take some

7:56

random text from the internet so we have

7:57

a sequence of words and and then we're

7:59

just going to try to predict uh the next

8:02

word based on previous words so let's

8:03

say we have uh we start with i here as

DESBLOQUEAR MÁS

Regístrate gratis para acceder a funciones premium

VISOR INTERACTIVO

Mira el video con subtítulos sincronizados, superposición ajustable y control total de la reproducción.

REGÍSTRATE GRATIS PARA DESBLOQUEAR

RESUMEN DE IA

Obtén un resumen instantáneo generado por IA del contenido del video, los puntos clave y las conclusiones.

REGÍSTRATE GRATIS PARA DESBLOQUEAR

TRADUCIR

Traduce la transcripción a más de 100 idiomas con un solo clic. Descarga en cualquier formato.

REGÍSTRATE GRATIS PARA DESBLOQUEAR

MAPA MENTAL

Visualiza la transcripción como un mapa mental interactivo. Comprende la estructura de un vistazo.

REGÍSTRATE GRATIS PARA DESBLOQUEAR

CHATEA CON LA TRANSCRIPCIÓN

Haz preguntas sobre el contenido del video. Obtén respuestas impulsadas por IA directamente desde la transcripción.

REGÍSTRATE GRATIS PARA DESBLOQUEAR

SACA MÁS PARTIDO A TUS TRANSCRIPCIONES

Regístrate gratis y desbloquea el visor interactivo, los resúmenes de IA, las traducciones, los mapas mentales y mucho más. No se requiere tarjeta de crédito.

    MIT 6.S087: Fou… - Transcripción Completa | YouTubeTranscript.dev