TRANSCRIPTEnglish

Vector store architectures beyond “just a vector DB”

8m 29s1,128 words204 segmentsEnglish

FULL TRANSCRIPT

0:00

Vector store architectures beyond just a

0:02

vector DB. Vectors are only half the

0:04

system. Metadata, filtering, scale, and

0:07

ops decide which architecture actually

0:09

works in production. One, the three AWS

0:12

native paths and why they exist. You'll

0:15

see these three over and over in exam

0:17

questions. Manage knowledge base KB.

0:20

Two, open search vector plus neural

0:22

plugin. Three, Amazon Aurora with PG

0:25

Vector plus Amazon DynamoB for metadata.

0:28

Each exists because one size does not

0:30

fit all. Two, manage knowledge base, the

0:33

easy button. What it is, a fully managed

0:35

rag setup, ingestion, chunking,

0:37

embeddings, storage, retrieval, minimal

0:40

configuration, minimal control. When AWS

0:43

wants you to choose it, fastest time to

0:45

value, simple documents, no custom

0:48

ranking logic, low ops burden, when not

0:51

to use it. Exam, you need custom

0:53

metadata filters. You need hybrid search

0:55

tuning. You need audit debug visibility.

0:58

You need custom chunking or reranking.

1:00

Memory hook. Knowledge base equals

1:03

automatic gearbox. You don't tune it,

1:05

you just drive.

1:07

How to three? Open search vector plus

1:09

neural plugin. Enterprise search engine.

1:12

What it gives you? Vector search and

1:14

keyword search. Hybrid search. BM25 plus

1:17

vector. Metadata filtering at scale.

1:20

Ranking. Fine grain tuning. This is the

1:23

most exam heavy option. Typical

1:25

architecture. Embedding stored as

1:27

vectors. Text fields indexed.

1:29

Traditionally filters on metadata. Date,

1:31

ro, tenant, region. Why AWS loves it?

1:35

Scales well. Handles complex queries.

1:37

Designed for search workloads. Memory

1:40

hook. Open search equals search brain.

1:43

Vectors plus keywords filters tuning.

1:46

The four Aurora PG vector relational

1:49

first thinking. What it is? A relational

1:51

database with vector support. Vectors

1:53

are just another column. When it shines,

1:55

you already use Aurora. Strong

1:57

relational constraints. Transactions

1:59

matter. Moderate data size. When it

2:01

struggles, very large vector sets, heavy

2:04

similarity search traffic, advanced

2:06

search features, memory hook. Aurora PG

2:09

vectoral first, vector second.

2:13

Now five, Dynamo DB, not a vector DB,

2:16

but crucial. Dynamob is not used for

2:18

similarity search. It is used for

2:20

metadata, access control, tenant

2:23

isolation, fast key value lookups,

2:25

common pattern, vectors, open search or

2:28

Aurora, metadata, Dynamo DB, join at

2:31

query time, memory hook, DynamoB equals

2:34

identity and control plane. Number six,

2:37

AWS static 1 vector architecture

2:39

edition. Static vector store choice

2:42

index schema metadata model search

2:44

strategy vector hybrid SQL plus one

2:46

incoming query user question. The

2:49

architecture stays fixed. Queries vary.

2:51

That's static plus one. Seven. How AWS

2:54

expects you to choose exam decision

2:56

table. Choose manage KB when simplicity

2:59

control low ops default rag is enough.

3:02

Choose open search when hybrid search

3:04

needed complex filtering. Ranking large

3:07

scale multi-tenant systems. Choose

3:09

Aurora PG vector when strong relational

3:12

needs moderate scale SQL joins matter

3:14

existing Aurora footprint. Use DynamoB

3:17

when metadata lookup permissions fast

3:20

keybased access. Now 8 classic exam

3:24

traps. Watch for these. Use DynamoB for

3:27

vector similarity. Knowledge base for

3:29

complex filtering. Aurora PG vector for

3:31

massive search workloads. One vector

3:33

store solves everything. AWS wants

3:36

intentional trade-offs, not blanket

3:37

answers. One memory story. Don't forget

3:40

this. The library system. Knowledge base

3:42

equals selfch checkckout kiosk. Fast,

3:45

simple, limited options. Open search

3:48

professional librarian knows keywords,

3:50

topics, cross references. Aurora PG

3:53

Vector filing cabinet with SQL labels.

3:56

Structured, orderly, not built for

3:57

crowds. Dynamo DB ID card system. Who

4:00

can access what? No single tool runs the

4:02

whole library. Exam compression rules

4:04

memorized. Simple rag. Manage KB search

4:07

plus scale filters. Open search. SQL

4:10

relations. Aurora PG vector metadata and

4:13

access Dynamo DB. If an answer ignores

4:16

metadata or filtering, it's incomplete.

4:19

What AWS is really testing, they are

4:21

asking open quote. Do you understand

4:23

that retrieval is a system not a

4:25

database? Close quote. If your answer

4:27

shows separation of concerns,

4:29

scalability thinking, metadata

4:31

awareness, you're choosing like AWS

4:33

expects. Below are three concrete

4:36

production style examples showing when

4:38

and why you'd pick managed KB, Open

4:40

Search or Aurora Page Vector plus

4:42

Dynamob. This is exactly how AWS expects

4:44

you to reason in the exam. Real example

4:47

one, fast rag for internal documents.

4:49

Manage knowledge base scenario. A

4:52

company wants an internal chatbot over

4:54

PDFs, policies, HR docs, word files,

4:57

procedures, requirements. Working in

4:59

days, not weeks. No custom ranking. No

5:02

complex filters. Low ops. Architecture.

5:05

Upload docs to S3. Use bedrock knowledge

5:08

base. Ask questions. Get grounded

5:10

answers. Why this is correct? KB handles

5:13

chunking, embedding, storage, retrieval,

5:16

minimal configuration, no infer. Why not

5:18

open search here? Overkill. More ops. No

5:20

extra benefit. Exam takeaway.

5:24

Simple rag. Low control needs. Managed

5:26

knowledge base. Memory hook KB equals

5:30

plug-and-play real example 2 large

5:33

enterprise search with filters open

5:34

search scenario a legal research

5:37

platform with millions of documents

5:39

queries like open quote find cases about

5:42

construction defects from NC at 2018 for

5:44

tenant X close quote needs semantic

5:47

similarity keyword matching metadata

5:49

filters date jurisdiction tenant

5:53

reranking architecture documents

5:55

embeddings store vectors text In Open

5:58

Search, use vector search, BM25, keyword

6:01

search, metadata filters, optional

6:03

reranker, access control via tenant

6:05

metadata. Why this is correct? Open

6:08

search supports hybrid search filters

6:09

run at scale designed for search

6:11

workloads. Why not manage KB? Limited

6:14

tuning, limited debugging, poor fit for

6:16

complex filters.

6:19

Exam takeaway

6:21

scales hybrid search filters. Open

6:24

search

6:25

memory hook. Open search equals

6:27

enterprise search brain. Real example

6:29

three. Transactional app with

6:31

embeddings. Aurora PG vector Dynamob.

6:35

Scenario. A SAS app where each user has

6:37

notes and comments. Notes need

6:39

similarity search. Find related notes.

6:42

Strong consistency. SQL joins with users

6:44

projects permissions. Data volume is

6:46

moderate not millions of vectors.

6:49

Architecture notes table in Amazon

6:51

Aurora. Content embedding vector project

6:53

ID. Use PG vector for similarity search.

6:56

Store permissions and fast lookups in

6:58

Amazon Dynamo DB. User allowed project

7:01

IDs. Query flow one user query

7:04

embedding. Two SQL query vector

7:06

similarity where project ID in. Three

7:09

DynamoB checks access. Why this is

7:12

correct? SQL first design transactions

7:14

matter moderate scale existing Aurora

7:16

footprint. Why not open search heavier

7:18

ops less natural for transactional

7:20

joins? Exam takeaway: Relational data

7:24

plus vectors. Aurora PG vector plus

7:26

Dynamob for metadata. Memory hook.

7:29

Aurora PG vector equals SQL app with

7:31

vector spice.

7:33

How AWS expects you to choose. Real

7:35

decision logic requirement correct

7:37

choice. Fastest rag low ops manage KB

7:41

hybrid search filter scale open search

7:44

joins transactions Aurora PG vector

7:47

metadata permissions dynamob

7:50

statics one real world framing static

7:53

vector store choice index schema

7:55

metadata model user query the system is

7:58

fixed queries vary that's static one and

8:00

retrieval architecture

8:03

one final memory story lock it in the

8:06

company library knowledge base Selfch

8:08

checkckout kiosk. Fast, simple, no

8:11

customization. Open search expert

8:13

librarian knows keywords, topics,

8:15

filters. Aurora PG vector equals filing

8:17

cabinet with SQL labels. Structure

8:19

transactional. Dynamob ID badge system.

8:22

Who can see what AWS wants you to pick

8:24

the right librarian, not bring a

8:26

bulldozer?

UNLOCK MORE

Sign up free to access premium features

INTERACTIVE VIEWER

Watch the video with synced subtitles, adjustable overlay, and full playback control.

SIGN UP FREE TO UNLOCK

AI SUMMARY

Get an instant AI-generated summary of the video content, key points, and takeaways.

SIGN UP FREE TO UNLOCK

TRANSLATE

Translate the transcript to 100+ languages with one click. Download in any format.

SIGN UP FREE TO UNLOCK

MIND MAP

Visualize the transcript as an interactive mind map. Understand structure at a glance.

SIGN UP FREE TO UNLOCK

CHAT WITH TRANSCRIPT

Ask questions about the video content. Get answers powered by AI directly from the transcript.

SIGN UP FREE TO UNLOCK

GET MORE FROM YOUR TRANSCRIPTS

Sign up for free and unlock interactive viewer, AI summaries, translations, mind maps, and more. No credit card required.