TRANSCRIPTEnglish

8 BILION DIGITAL CLONES

12m 56s2,566 words369 segmentsEnglish

FULL TRANSCRIPT

0:00

A few years ago, there was this

0:01

experiment out of Stanford to see what

0:03

happens if you populate a village with

0:05

large language models. Can you simulate

0:07

what happens? Can it sort of simulate

0:09

what happens in the real world? It was

0:10

by Jun Park, Smallville, aka interactive

0:14

similacra. Absolutely loved covering

0:16

that story. Highly, highly fascinating

0:18

paper and now it's back and bigger than

0:21

ever. So, Jun Park, the the Stanford

0:23

researcher, so he's back with something

0:25

that he's calling simile. And the basic

0:28

concept is to take this idea of digital

0:30

twins and create entire societies. So

0:33

let's say you take entire demographics

0:34

based on transcripts, transaction logs,

0:37

scientific data, whatever. So you create

0:39

societies, cities, villages, whatever.

0:41

Then you run the simulation. And the

0:43

goal here is to be able to ask certain

0:45

specific questions about what happens.

0:47

What happens if we raise or lower taxes?

0:50

What happens if there's a new marketing

0:52

campaign for a specific product? How do

0:54

people react to this news breaking out?

0:57

Really anything you want, but

0:59

specifically targeting kind of like the

1:00

social interactions. How do certain news

1:03

flow through the community? How do

1:04

people respond? And this isn't like a

1:07

little project. This thing is massive

1:10

and has some pretty big heavyweights on

1:12

board. So there's a few angel investors

1:14

including Andre Karpathy. So he's the

1:16

co-founder of OpenAI. He was a former

1:19

director of AI at Tesla, a big name in

1:21

the industry. You have Faith Lee who is

1:24

the the godmother of AI. She's the

1:26

co-director of Stanford's human centered

1:28

AI institute. We also have Adam

1:30

D'Angelo. So you might remember him from

1:32

that whole OpenAI fiasco where Sam Alman

1:35

got fired. He was one of the board

1:36

members. He's still on the board of

1:38

OpenAI. He's the CEO of Quora. Then we

1:40

have GMO Raj who's the CEO of Versell.

1:43

Huge company widely used. Recently I was

1:46

using my AI agents OpenCloud to build

1:48

something with it. I suggested they use

1:50

WordPress. They were like, "No, dude.

1:51

We're not using WordPress. We're using

1:53

Versell." Is it Versel? I don't even

1:54

know how to pronounce it, but they told

1:56

me that's the way to go. And I went

1:57

along with it. Very happy. You know,

1:59

five out of five stars would let my AI

2:02

agents use it again. And it's called

2:03

Bilski, Adobe's chief strategy officer

2:05

and founder of Behance. Okay, but what's

2:08

the big deal here? So, the original

2:10

Similac paper was kind of mind-blowing

2:13

because it was very very early in sort

2:15

of like the the chat GPT story. I think

2:17

they were using chat GPT or GBT 3.5

2:21

Turbo at the time to simulate all the

2:24

agents in the village. I believe there

2:25

were 25 agents or you know little

2:26

characters and they had those cutesy

2:28

pixel art animations and each of them

2:31

had a backstory personality. They had

2:33

jobs or they went to school. They were

2:36

married to people. They had kids etc.

2:37

And they had a little schedule too. So

2:39

they kind of woke up oh 8 a.m. I got to

2:41

go to work or I got to go to school etc.

2:43

So they literally lived out their entire

2:46

lives, did their chores, interact with

2:49

their loved ones, etc. And at the time

2:51

the question was very simple. What

2:53

happens if we give one of them kind of a

2:57

inspirational idea? So imagine like

2:59

almost like it's some sort of a divine

3:00

inspiration where God shows up in your

3:03

sleep and says do this and you're like

3:05

ah I got to go do this. So they did that

3:07

to one of the characters. I think

3:09

Isabella was her name. This was years

3:11

ago. So, I apologize if I miss some

3:12

details, but they told her, "You shall

3:15

create a a Valentine's Day party for

3:18

everyone in the city." Interestingly, I

3:19

just realized that they're launching it

3:21

a day before Valentine's Day. And in the

3:23

past, that whole thing was about

3:25

organizing a Valentine's Day party. But

3:27

the point was to see how well this

3:30

community would sort of be able to

3:32

create this party because again, they

3:34

only notified one person. So this one

3:36

person had to first and foremost tell

3:39

everybody else, convince them to have

3:41

it. Then they had to plan the party,

3:43

they had to organize the party, etc.,

3:46

etc., etc. So on day one or day zero,

3:49

whatever you want to call it, it's only

3:50

Isabella knew that that she had this

3:52

idea. No one else was thinking about it.

3:54

So she went, she talked to her closest

3:56

friends and she told her, "Hey, let's do

3:58

this party." And some of them liked it

4:00

and some of them didn't. But the news

4:01

kind of percolated throughout their

4:03

little village, their their little

4:04

society. To fast forward to the end of

4:06

it, like they did have the Valentine

4:08

party. Not every single person who came

4:12

was invited by Sabella, the original

4:14

person. So some of it kind of

4:15

percolated. So she would tell person A,

4:17

person A would tell person B, and then

4:19

person B would would show up. So there

4:21

was this kind of like ways of

4:22

information traveling through people

4:24

similar to how it does in the real life

4:26

in the real world. And some people

4:28

wanted to show up, some people forgot to

4:30

show up, some people did not want to

4:32

show up and didn't show up. So it very

4:33

much simulated how, at least in that

4:36

sort of small village, small society,

4:37

how things could have went. And one

4:41

highly, highly interesting thing about

4:42

this paper, looking back at it now, is

4:45

they figured out a lot of ways and a lot

4:47

of scaffolding to create around these

4:49

large language models that were highly

4:51

highly effective. They had, for example,

4:52

a memory stream. and they would organize

4:54

their memories in terms of how relevant

4:56

they were in terms of like the recency

4:58

etc. So over time instead of just

5:00

forgetting everything they kind of had

5:02

their most core memories and that that

5:04

kind of memory pool was added to over

5:06

time so they just didn't forget

5:08

everything all of a sudden. The

5:10

interesting thing about that is now with

5:12

AI agents like OpenClaw I mean we're

5:14

seeing a lot of the same ideas being

5:16

incorporated to make these agents highly

5:19

highly highly capable. So I think a lot

5:21

of us expected more to come out of this.

5:25

We knew there was going to be more and

5:26

more projects, bigger projects coming

5:27

out of this because the whole idea

5:29

seemed number one useful for predicting

5:30

human behavior for simulating it and

5:33

just I think it captured a lot of

5:34

people's imaginations and we didn't hear

5:37

anything for quite some time and finally

5:39

today I think they're coming out of

5:41

stealth mode and talking about it and

5:43

apparently they already had a $100

5:45

million seed round. So this thing has

5:48

some serious people behind it. It's got

5:50

some serious money behind it and

5:52

apparently it already has some large

5:55

enterprise clients that are behind it

5:57

that are on board. CVS Health, a

6:00

pharmacy store here in the United States

6:02

as well as Telra. They provide mobile

6:04

coverage, mobile phones, internet, etc.

6:06

And so some of the things that these

6:07

companies might want to do is for

6:09

example have market research done to see

6:12

how a group of let's say a thousand

6:13

people might react to a brand new

6:15

marketing campaign or product testing UI

6:18

testing etc. Jun Park reported that this

6:22

has been pretty accurate in terms of

6:24

predicting the analyst questions during

6:27

simulated earning calls. So the models

6:29

correctly predict eight out of 10 of

6:32

these earning calls that are asked by

6:33

analysts. So, if you think about it,

6:35

that could be extremely extremely

6:37

useful. You're doing an earning call.

6:38

It's live. People are not going to be

6:40

asking you tough questions, but you can

6:41

run the simulation and it tells you

6:43

like, here's most likely what they're

6:45

going to ask. So, you can prepare for

6:46

it. This can also be used for social

6:48

science. How will people respond to new

6:51

health scares or economic shocks, for

6:54

example. I think they should rerun 2020

6:56

and and see like is the whole toilet

6:59

paper being gone everywhere. Was that

7:01

predictable? Like, was it was it

7:02

obvious? because I was aware of that

7:04

whole thing happening very very early on

7:06

but somehow the whole toilet paper thing

7:08

was was never on my bingo card. So why

7:11

is this such a big deal? Now if you

7:13

think about it this kind of might mark a

7:16

transition from big data to big

7:19

simulation meaning that in the past

7:21

companies the more data they had that

7:23

was kind of like the gold that was the

7:24

the oil digital oil because you can mine

7:27

that data to prepare for certain events

7:29

to forecast etc. So having that data

7:32

stored somewhere and then kind of

7:33

parsing through it and trying to extract

7:35

insights was a big big deal. If these

7:38

simulations get large enough and

7:39

accurate enough that might not be as

7:42

important like that that realworld data

7:44

stored somewhere might not be as

7:46

important. Instead it's going to

7:47

transition to running these simulations

7:50

to try to extract data like that virtual

7:52

simulation data from there. So instead

7:55

of going out there and interviewing

7:57

10,000 customers to try to see what they

8:00

think and then you know like getting

8:01

that data, you run a simulation with

8:03

100,000 customers and you kind of figure

8:06

out what they think based on that

8:07

simulation or how they would react to a

8:09

new brand new marketing campaign. The

8:11

other big shift here might be the fact

8:13

that you know a lot of startups there's

8:15

this thing that's referred to sometimes

8:17

as the innovation tax, right? You might

8:19

have heard of it as the pioneers have

8:21

arrows in their backs. Basically the

8:23

whole idea that when you're the first to

8:25

market with something, when you have

8:26

some innovative idea, there's a cost to

8:28

it because if you fail, you pay the

8:30

price because a failure is expensive.

8:32

But yet, if you figure something out,

8:34

then other people can come in and sort

8:36

of copy what you're doing, etc. So, the

8:38

first movers, you know, there's a first

8:40

mover advantage, but there's also an

8:42

innovation tax, and the two kind of

8:44

balance each other off. But imagine if

8:46

there wasn't an innovation tax. What if

8:49

a startup or a brand new company, what

8:51

if it could run a thousand simulations

8:54

about how something will unfold, a new

8:57

product or anything like that and it can

8:58

run it in those digital simulation

9:01

sandboxes, right? Gather the data, see,

9:03

okay, it seems like this is the wing

9:05

approach. This is kind of like the more

9:07

dangerous approaches, etc. So, what if

9:09

the cost of running those thousand

9:11

simulations are equivalent to the

9:13

expense of actually executing one of

9:16

those in the real world? That means you

9:18

can try a thousand times for that same

9:20

price. That really cuts down on that

9:21

innovation tax. You can fail 999

9:26

times in that simulation as long as you

9:27

find that that one sort of winning

9:29

outcome. Now, of course, a lot of this

9:31

is going to be a little bit more

9:32

statistics based, right? So maybe the

9:34

simulation is going to be accurate only

9:35

85% of the time or whatever. I say 85%

9:38

because in that statement by Jun Sun

9:40

Park, the the researcher behind this,

9:42

that's what they found in relation to

9:43

those uh earnings calls, right? it was

9:45

85% of the time it was accurate. And

9:48

again, it's only going to keep getting

9:50

better and better. But let's assume it's

9:51

just 85%. If you can run simulations

9:54

that are 85% accurate, even that would

9:57

be incredibly beneficial. But here's

10:00

kind of what I think is the big thing,

10:02

the big deal that might have a outsiz

10:05

effect. A lot of our data statistics,

10:07

they they kind of they regress to the

10:08

mean. They're kind of trying to find the

10:11

average person. What does the average

10:13

person think about this phone, right?

10:15

How do they rate it? They rate it 8 out

10:17

of 10 or whatever. But the reality is

10:20

across that average, there's a lot of

10:22

different people for a lot of different

10:23

opinions. Some of them have an outsized

10:26

effect on the market. And just looking

10:28

at the average doesn't really

10:29

necessarily predict how people will

10:32

behave. What if 99% of the people are

10:34

okay with something and then 1% of the

10:37

people that get really triggered by a

10:39

particular message that you're putting

10:41

out there or some feature that your

10:43

product has, they go, you know, declare

10:46

a war and make lots of noise about it.

10:48

The rest of the people that are, you

10:49

know, they're okay. They're look warm

10:51

with the product. They like it, but they

10:52

don't love it, but they don't hate it.

10:54

But they do see this one group of people

10:55

that are damaged by the product or

10:57

triggered. What whatever happens, right?

10:59

we're may be able to simulate those

11:02

weird idiosyncratic things that that

11:04

kind of happen out there in the real

11:05

world cuz again, you know, if you're

11:07

looking at averages, on average

11:08

everybody's fine, right? Everybody sort

11:09

of likes it. No problems. But that

11:12

minority of people might have an

11:13

outsized effect. And this would allow us

11:15

to capture those weird trends and and

11:18

weird eentricities that might happen

11:20

that that we can see in a simulation

11:22

that we can't if we're running some

11:25

statistical analysis on a bunch of data.

11:27

And of course, this could be huge for

11:29

stock market analysis. You might

11:31

simulate CEOs and market traders and a

11:34

lot of people involved like leadership

11:36

teams and companies to see how they

11:38

react to a market crash or a

11:40

competitor's moves and see if you can

11:43

predict some things before they occur.

11:45

Certainly, there were conversations in

11:46

my life that I'd love to be able to run

11:49

through something like this to to

11:50

hopefully get a glimpse into how they

11:52

how might they go. like if we have to

11:54

give this person some tough news, what

11:56

are some possible sort of outcomes of

11:59

this happening? What might they say? How

12:01

may I react? And would that trigger

12:03

anything? Anyways, very excited to see

12:06

more coming out of this company. Similey

12:08

and June, of course, I'm so happy that

12:10

he's having success. Obviously,

12:11

incredibly intelligent person, one of

12:13

the first really mind-blowing papers

12:15

utilizing Chad GPT. Absolutely loved

12:18

everything that he did there and

12:20

expecting to see a lot more from this.

12:22

They're on Twitter x simile company at

12:25

simile_ai.

12:26

So I think this company is going to

12:28

answer a lot of questions that people

12:30

have and it might even raise a new

12:32

question so to speak and that is if

12:34

running simulations to get data if it's

12:36

so incredible and I'm sure the data is

12:38

more valuable the more accurate a

12:40

simulation is then what's the chance

12:42

that this is a simulation and we're all

12:45

just here stuck reacting to some

12:48

marketing campaign. Something to think

12:49

about. If you made this far, thank you

12:51

so much for watching. My name is Wes

12:52

Roth. I'll see you in the

UNLOCK MORE

Sign up free to access premium features

INTERACTIVE VIEWER

Watch the video with synced subtitles, adjustable overlay, and full playback control.

SIGN UP FREE TO UNLOCK

AI SUMMARY

Get an instant AI-generated summary of the video content, key points, and takeaways.

SIGN UP FREE TO UNLOCK

TRANSLATE

Translate the transcript to 100+ languages with one click. Download in any format.

SIGN UP FREE TO UNLOCK

MIND MAP

Visualize the transcript as an interactive mind map. Understand structure at a glance.

SIGN UP FREE TO UNLOCK

CHAT WITH TRANSCRIPT

Ask questions about the video content. Get answers powered by AI directly from the transcript.

SIGN UP FREE TO UNLOCK

GET MORE FROM YOUR TRANSCRIPTS

Sign up for free and unlock interactive viewer, AI summaries, translations, mind maps, and more. No credit card required.