TRANSCRIPTIONEnglish

MIT 6.S087: Foundation Models & Generative AI. CHAT-GPT & LLMs

1h 5m 6s11,635 mots1,658 segmentsEnglish

TRANSCRIPTION COMPLÈTE

0:00

all

0:01

right welcome to the third lecture on

0:06

Foundation mulative AI So today we're

0:09

going to cover chat

0:11

GPT um and um right I mean I think for a

0:16

lot of people chat GP was

0:18

the the tool or the the AI that really

0:23

made people understand this is different

0:25

now we're able to do things we weren't

0:27

able to do before and and definitely uh

0:30

created some kind of hype uh so

0:33

hopefully after this lecture you'll you

0:35

understand kind of the basic idea and

0:37

also somehow understand the BET right

0:39

the bet that open Ai and Ilia the head

0:42

researcher did in terms of what actually

0:44

would lead to CHP and how in hindsight

0:47

it might be quite I mean easy but it was

0:52

a really daring bad not obvious at all

0:55

at the time that this would actually

0:56

work out

0:58

um so should be be a lot of fun and just

1:01

to quickly go through our course

1:02

schedule as well a little bit right so

1:04

today is January 16 uh and next time

1:07

we'll talk about stable diffusion image

1:10

generation and then we'll talk about

1:12

emerging Foundation models basically

1:14

Foundation models generative AI in the

1:16

commercial space H we'll have two guest

1:19

speakers and then we'll end with the

1:21

lecture on AI ethics and regulation as

1:24

well as a

1:26

panel okay so what have we talked about

1:28

before we started off

1:30

H with an introduction a short high

1:33

level intuitive answer to what is

1:35

foundation M generative

1:36

AI we went a little bit on a

1:38

philosophical digression and asked about

1:41

how's the world structured because that

1:43

allows us to think about how we should

1:44

learn in the world then we on the second

1:47

lecture went through all the different

1:48

algorithms um and yeah today we'll we'll

1:52

dive in more specifically into chpt and

1:54

kind of uh pull everything together um

1:58

and to reiterate right so what do we do

2:01

in uh Foundation models geni well we

2:04

apply this self-supervised learning

2:06

where we learn without uh label data so

2:09

we can we can get you know as much data

2:12

as we want because there's no human

2:13

being in the loop so there's no limit

2:15

how much we can scale this up and and

2:18

what we get from this you know by

2:19

learning from observation and learning

2:21

from the data directly is a very

2:23

contextual and relational understanding

2:25

of meaning and we gave this example

2:27

before about you know from a supervised

2:30

learning perspective you learn what a

2:32

dog is from seeing you know labeled uh

2:36

examples of dogs and in reinforcement

2:39

learning you focus on optimizing certain

2:41

goals and you understand a dog in

2:43

relation to how it makes you happy or

2:45

fulfilled in some sense or optimizing

2:47

your goals but in self supervised

2:49

learning right it's the foundational

2:50

technology behind uh Foundation models

2:53

you learn from observing dogs in

2:54

different context and you get a very

2:56

relational definition of a dog so it's

2:59

something that's walk by an owner with a

3:00

leash it has an anistic R with cats it

3:03

chases fris with oone right this is your

3:06

definition of what a dog is and today

3:09

we'll you talk about something that's

3:11

extremely engineering heavy in you know

3:15

chat GPT uh relies on a lot of tricks

3:19

and Engineering insights and

3:21

breakthroughs that we're not going to

3:22

cover and I think still though you know

3:26

like it's like talking about a car you

3:28

can understand the high level

3:29

perspective of a car and get some

3:31

insights how to work how it works and

3:32

how it's going to be useful for you

3:34

without getting into all the engineering

3:36

details but of course in real life those

3:39

engineering details really really

3:40

matters and are very very hard to get

3:41

right and that's something that we won't

3:43

really dive into in this lecture because

3:46

that's just when you bring something up

3:48

certain scale and you have to paralyze a

3:49

lot of machines Etc and think about high

3:52

parameters it's a whole science so it's

3:54

not trival at all but it's kind of hard

3:57

uh to teach in a course like this and

3:59

and you have to learn by just actually

4:01

building this

4:03

stuff

4:05

um okay so

4:09

um also a little bit of philosophizing

4:12

in this uh class as

4:15

well

4:17

um I think that again like we talked

4:21

about a little bit of a theme here right

4:22

is that the why this new AI is so

4:25

powerful is because it doesn't Force

4:28

things to comply to Simple Rules right

4:32

it kind of abandons our ability to

4:34

understand and compress what we're

4:35

seeing and deals with that chaos

4:37

directly that's why AI is so powerful

4:38

and so

4:39

humanlike um so also like when I talk

4:42

about this in CHP we try to make very

4:45

high level um statement but of course

4:49

the nuances matters and I think it's

4:51

quite interesting uh I took this quote

4:53

from a general from the 18 and

4:58

1700s and he says this uh quote that P

5:02

Theory which sets itself in opposition

5:04

to the mind and what he meant was that

5:06

he's a general so he fights in battles

5:09

and War and at the time people loved to

5:13

come up and theorize around War like we

5:15

should have certain rules and how

5:16

soldiers should behave in fighting and

5:18

stuff like that but he's like well I've

5:20

been in War uh and Wars don't comply to

5:24

rules first off so you know everybody

5:28

has a plan before they get hit in face

5:29

basically so you know as people start

5:31

shooting at you and you have this fog of

5:33

War of you don't know what's going on

5:35

there's no simple rules to help you

5:37

there and also what he says this in

5:39

terms of the mind he says like well

5:41

actually he's realized by working with

5:43

soldiers that soldiers and human beings

5:45

our mind we're not good at acting

5:49

according to rules that we try to

5:50

memorize we're very intuitive and very

5:52

kind of quick to react to things by our

5:55

intuition that's what really really

5:56

matters and that's what we're strong at

5:57

so if you force a soldier's well Al try

5:59

to memorize a lot of rules and that's

6:01

how it should act in a battle you're

6:03

kind of screwed and very limited in what

6:04

you can do uh which also is something

6:07

that I think AI uh in a new type of AI

6:12

leverages okay so chat

6:16

GPT um right this is a really amazing

6:20

breakthrough that uh has some very

6:22

humanlike Mastery of language that we

6:24

can communicate that can basically solve

6:27

a really wide array of tasks for us

6:30

anything that can be phrased in terms of

6:32

text language it can it can basically

6:35

solve and now as well when with gp4 ET

6:38

becomes uh it's able to handle multi

6:40

modalities but it's it's extremely

6:42

powerful so let's try to break this

6:45

apart well first off what does this name

6:48

actually stand for well the chat part is

6:51

obvious it stands for chat and then GPT

6:53

stands for generative pre-trained

6:56

Transformer and this is a I mean a good

6:59

description of what this uh actually is

7:02

um and I think also if you look at the

7:06

the two different three different

7:08

concepts here they're also almost

7:10

corresponding length in terms of how

7:12

important and influential they are in

7:14

making chat GPT work so chat part we

7:18

we'll cover last it's the kind of the

7:20

least important one in some sense H the

7:22

Genty pre-trained is the self supervised

7:25

step of how you train this and arrive at

7:26

this uh model and then the Transformer

7:29

is the basically the engine behind it in

7:31

some sense

7:33

and so let's start with this generative

7:37

pre-train what does it mean how do we

7:39

pre-train this model and that's

7:41

basically where openi spent 99% of the

7:45

compute was to do this pre-training step

7:47

so it's it's very very

7:50

important okay so what we're going to do

7:52

is that we're going to uh just take some

7:56

random text from the internet so we have

7:57

a sequence of words and and then we're

7:59

just going to try to predict uh the next

8:02

word based on previous words so let's

8:03

say we have uh we start with i here as

DÉBLOQUER PLUS

Inscrivez-vous gratuitement pour accéder aux fonctionnalités premium

VISUALISEUR INTERACTIF

Regardez la vidéo avec des sous-titres synchronisés, une superposition réglable et un contrôle total de la lecture.

INSCRIVEZ-VOUS GRATUITEMENT POUR DÉBLOQUER

RÉSUMÉ IA

Obtenez un résumé instantané généré par l'IA du contenu de la vidéo, des points clés et des principaux enseignements.

INSCRIVEZ-VOUS GRATUITEMENT POUR DÉBLOQUER

TRADUIRE

Traduisez la transcription dans plus de 100 langues en un seul clic. Téléchargez dans n'importe quel format.

INSCRIVEZ-VOUS GRATUITEMENT POUR DÉBLOQUER

CARTE MENTALE

Visualisez la transcription sous forme de carte mentale interactive. Comprenez la structure en un coup d'œil.

INSCRIVEZ-VOUS GRATUITEMENT POUR DÉBLOQUER

DISCUTER AVEC LA TRANSCRIPTION

Posez des questions sur le contenu de la vidéo. Obtenez des réponses alimentées par l'IA directement à partir de la transcription.

INSCRIVEZ-VOUS GRATUITEMENT POUR DÉBLOQUER

TIREZ LE MEILLEUR PARTI DE VOS TRANSCRIPTIONS

Inscrivez-vous gratuitement et débloquez la visionneuse interactive, les résumés IA, les traductions, les cartes mentales, et plus encore. Aucune carte de crédit requise.

    MIT 6.S087: Fou… - Transcription Complète | YouTubeTranscript.dev