TRANSCRIPTEnglish

I Wasn't Ready For What This Does

9m 38s1,937 words300 segmentsEnglish

FULL TRANSCRIPT

0:00

With how powerful generative AI has

0:02

become, new tools are constantly

0:03

emerging. And amongst all, Google has

0:05

really stepped up its game since Gemini

0:07

3 came out. Because of this, you've

0:08

probably seen people on X oneshotting

0:10

these amazing landing pages and saying

0:12

that the model is a gamecher. But

0:14

they're lying. The truth is they need to

0:16

use multiple tools to build these sites.

0:18

And Google has been going crazy with

0:19

their experimental tools, which are

0:21

powered by Gemini 3 and Nano Banana. But

0:23

Google doesn't offer a way to use all of

0:25

them together. So, you're going to need

0:27

a complete workflow that combines all

0:28

these tools. We've got a lot to talk

0:30

about today as I wasn't really expecting

0:32

them to work so well together. The

0:34

animations that you see on those sites

0:36

are just a series of frames. But if you

0:37

use AI tools to generate these frames,

0:40

you don't get consistent results. Google

0:41

solves this by releasing an experimental

0:43

tool called Whisk, which is particularly

0:45

designed for assets generation. It is

0:47

powered by Nano Banana for image

0:49

generation. I often use Whisk to

0:51

generate sequences of images for hero

0:53

sections. I provide prompts in simple

0:55

words, update the details step by step

0:57

for each sequence, and then use the

0:59

resulting images in the hero sections.

1:01

Using this whole process, I created this

1:03

landing page and was able to implement

1:05

this cool animated effect with the

1:06

camera. To start, we're going to

1:08

generate just the first and last frames

1:10

of a sequence and then create an

1:11

animation using those two key frames for

1:13

the hero section. But if you just start

1:15

prompting it without any structure, then

1:17

the key frames you want won't have the

1:19

same structure continuity. For this

1:21

purpose, you have to clearly define the

1:22

subject, the visual intent, and the

1:24

level of detail you want in the image.

1:26

Whisk uses a subject, scene, and style

1:28

framework to guide image generation,

1:30

allowing you to combine them into a

1:32

single new visual. That's why I included

1:34

all the details on what kind of camera I

1:36

wanted, how I wanted the reflections on

1:38

the lens to look, the resolution, and

1:40

the depth of the image, and it created

1:42

the visual exactly as the prompt

1:44

outlined. The generated image will not

1:46

always meet your expectations, in which

1:48

case you just have to mention the change

1:49

in simple words and it incorporates the

1:52

changes into a new visual. What I like

1:53

most about Whisk is that you don't have

1:55

to write long detailed prompts to get

1:57

great results. This is because it uses

1:59

Gemini 3 as a middle layer which writes

2:01

the detailed prompt on top of your

2:03

simple words leading to solid visuals.

2:05

Now, it raises the question of why I

2:07

chose Whisk. While Nano Banana requires

2:09

extensive text prompts, and Google's Mix

2:11

Board is designed for mood boards,

2:13

neither is optimized for fast,

2:15

controlled image remixing. Whisk's core

2:17

strength lies in combining reference

2:19

images, allowing you to generate images

2:20

with better control. Once I had my first

2:22

frame, I wanted my last frame to be a

2:24

side angle of the camera with the lens

2:26

taken apart to show the components. So,

2:28

I prompted it to generate a technical

2:30

cutaway, specifying how it should layer

2:32

the internal lenses and how the

2:34

background should appear. It didn't get

2:35

it right on the first try. It broke down

2:37

the internal circuitry too which I

2:39

didn't want it to show. As I said

2:40

earlier, you just mentioned the change

2:42

you need to make. So I instructed it to

2:44

only show the lens layering after which

2:46

it successfully generated the image

2:48

without showing the internal circuitry.

2:50

Now whisk also supports animations using

2:52

the VO model. But Whisk's animations

2:54

focus on animating one image rather than

2:56

being able to connect multiple key

2:58

frames together. That's why I used

2:59

another tool called Google Flow. Flow

3:01

uses Gemini for planning the story,

3:03

Google's image models to design steady

3:05

characters, and VO to create video

3:07

content. I had already created my

3:08

starting and ending frames for the

3:10

camera animation, and now I wanted to

3:12

create a transition in them. So, I used

3:13

the frame to video option and provided

3:15

my frames. In order to ensure a smooth

3:17

transition, you need to mention in the

3:19

prompt how the starting frame leads to

3:21

the ending because it provides the model

3:23

with a logical path. Your prompt should

3:24

include how you want the animation to

3:26

flow, how the starting frame should

3:28

transit into the ending frame, and the

3:29

animation style, as these details ensure

3:31

the motion is intentional. Sometimes

3:33

these models tend to make mistakes with

3:35

complex tasks, so it didn't get my

3:37

animation right the first time. The

3:39

generated video got both the spin

3:41

direction and the ending frame

3:42

completely wrong, making the transition

3:44

awkward. The fix was simply reprompting

3:46

with some the necessary changes to

3:48

ensure the animation was correct. Just

3:49

like how I prompted it to change the

3:51

direction of the spinning of the camera

3:53

for smoother transition after which it

3:55

produced a version I wanted which I

3:56

downloaded for use in my project. Now

3:58

the video generation is not available on

4:00

the free tier unlimitedly because video

4:03

generation models are costly. It only

4:05

provides 180 monthly credits depending

4:07

on the region. Since each video

4:09

generation with VO3.1 uses 20 credits,

4:12

you get up to nine videos per month.

4:14

Since the videos generated by Flow are

4:16

in MP4 format and can't be used directly

4:19

in hero sections because they are harder

4:21

to map through scroll animations in. For

4:23

this reason, I converted them to WEBP

4:25

using a free online converter. I

4:26

uploaded the MP4 video, set the

4:28

conversion settings to produce the best

4:30

quality animated WEBP and it converted

4:32

to a WEBP format which I downloaded for

4:34

use in my project. Choosing WEBP is

4:36

important because with this it is easier

4:38

to map scroll interactions since on the

4:41

web this format is treated as an image

4:43

which doesn't require a media player

4:45

wrapper like other formats do. These

4:47

files are more compact and deliver

4:48

improved performance making them ideal

4:50

for short form animated content. I added

4:52

the converted WEBP file to the public

4:54

directory of my newly initialized Nex.js

4:57

project because this is where all the

4:59

assets reside in the project. Now once

5:01

we had our animation ready, I wanted it

5:03

to add it to the hero section in my

5:04

landing page. I generally prompt Claude

5:06

in XML format because their models are

5:08

optimized for understanding XML, letting

5:11

them parse structured intent more

5:12

reliably and reason over each section

5:14

independently. The prompt I gave Claude

5:16

for adding the animation included

5:18

context on what I wanted to build, where

5:20

the assets for the animation are

5:22

located, and how the scroll through

5:23

animation should work and our goal in

5:25

the context tag. I included all the

5:27

requirements in the requirement tags,

5:29

described how the animation should

5:30

behave in animation behavior tags, and

5:33

specified the implementation details,

5:35

constraints, and required output

5:36

directly in the prompt within their

5:38

respective tags. When I gave Claude this

5:40

prompt, it automatically implemented the

5:42

animation exactly as I wanted. Even

5:44

though our hero section looked good, the

5:46

rest of the components looked like the

5:47

usual generic websites AI tends to

5:50

generate. This is because we were

5:51

expecting highquality results from

5:53

vanilla CSS instead of relying on

5:55

existing beautiful component libraries.

5:57

There are multiple UI libraries, each

5:59

with its own specialized style and

6:01

design themes, but you have to choose

6:03

the library that suits your project

6:05

style best. For my camera landing page,

6:07

I was going for an Apple style UI. And

6:09

the closest library to that idea is Hero

6:11

UI. It's built on top of Tailwind CSS

6:14

and relies on framer motion to bring its

6:16

components to life across the website.

6:18

The library supports most commonly used

6:20

frameworks like Nex.js, Vit and Laravel.

6:22

So using it with my existing Nex.js

6:24

setup was easy. There are two methods of

6:26

installation. Either you install it

6:28

projectwide with all components

6:30

available via the install command or you

6:32

install individual components as needed,

6:34

which is what I did with Claude. I

6:35

prompted Claude to replace the existing

6:37

components with Hero UI components and

6:39

the website was significantly improved,

6:41

giving the site a much more professional

6:43

look and feel. Users decide whether to

6:45

stay on a landing page within a few

6:47

seconds by looking at how engaging the

6:49

UI is. Motion helps guide the visitors

6:51

attention to the features we want them

6:53

to notice, ensuring higher user

6:55

retention. Implementing animations from

6:57

scratch using vanilla JavaScript can be

6:59

challenging, so I rely on existing

7:00

libraries to simplify the process. For

7:02

this project, I used motion.dev, a free

7:05

and open-source library that offers a

7:07

wide range of readytouse animations.

7:09

Normally, animations would require

7:10

manually syncing DOM updates with

7:12

animation timings. However, motion.dev

7:15

abstracts this logic by handling DOM

7:16

updates internally. It automatically

7:18

updates components as the user scrolls,

7:21

so animations play smoothly without the

7:23

need to manually track scroll positions.

7:25

This library uses motion components

7:27

instead of standard ones. These

7:29

components have start and end states

7:31

defined in the props and the library

7:33

handles the transitional logic between

7:34

them automatically. For our landing

7:36

page, I prompted Claude to implement

7:38

parallax and scroll animations using the

7:40

library. As a result, the user

7:42

experience improved by guiding attention

7:44

toward the key sections of the page.

7:46

Describing how the sections of a page

7:48

should look is a tedious process. It is

7:50

better to get inspiration from existing

7:52

galleries where people post their

7:54

creations. I use 21st.dev, a platform

7:56

that offers inspiration for a variety of

7:58

UI components built by multiple

8:00

designers. The components are built on

8:02

top of popular UI libraries like

8:04

Aceternity UI, Prism UI, Coconut UI,

8:07

Magic UI, and many others. While looking

8:09

for ideas, I came across this call

8:11

toaction section that would look great

8:13

on my landing page. The part I like best

8:15

about 21st.dev is that for any component

8:18

I want to use, I can just copy a prompt

8:20

specifically tailored for AI coding

8:22

agents. I don't need to guide Claude

8:24

myself. The prompt is extensively

8:26

structured, including project

8:27

requirements such as shad CN and

8:29

TypeScript support. It provides code

8:31

with instructions for the coding agent

8:33

to paste directly into the components

8:35

directory. It includes all necessary

8:36

dependency code and the paths where it

8:38

should be added and it lists the

8:40

required npm packages to install. It

8:42

also includes an implementation guide

8:44

for your AI agent detailing all the

8:46

steps needed to integrate the component

8:48

directly into your application and how

8:50

the agent should tailor it to the

8:51

specific project's needs. I gave this

8:53

prompt to Claude and it integrated the

8:55

exact same call to action section for

8:58

which I had copied the prompt. It also

8:59

added motion from the motion library we

9:01

had installed. Even though I did not

9:03

explicitly mention adding motion

9:05

anywhere, I also got the footer from

9:07

21st. Even though the demo footer

9:09

included icons for GitHub and Twitter,

9:12

when I gave Claude the copied prompt, it

9:13

omitted the GitHub icon since it wasn't

9:16

relevant to our project. It customized

9:17

the footer to include only the icons

9:19

related to the camera product site,

9:21

creating a footer that perfectly fit our

9:23

landing page. That brings us to the end

9:25

of this video. If you'd like to support

9:27

the channel and help us keep making

9:28

videos like this, you can do so by using

9:30

the super thanks button below. As

9:32

always, thank you for watching and I'll

9:34

see you in the next one.

UNLOCK MORE

Sign up free to access premium features

INTERACTIVE VIEWER

Watch the video with synced subtitles, adjustable overlay, and full playback control.

SIGN UP FREE TO UNLOCK

AI SUMMARY

Get an instant AI-generated summary of the video content, key points, and takeaways.

SIGN UP FREE TO UNLOCK

TRANSLATE

Translate the transcript to 100+ languages with one click. Download in any format.

SIGN UP FREE TO UNLOCK

MIND MAP

Visualize the transcript as an interactive mind map. Understand structure at a glance.

SIGN UP FREE TO UNLOCK

CHAT WITH TRANSCRIPT

Ask questions about the video content. Get answers powered by AI directly from the transcript.

SIGN UP FREE TO UNLOCK

GET MORE FROM YOUR TRANSCRIPTS

Sign up for free and unlock interactive viewer, AI summaries, translations, mind maps, and more. No credit card required.