文本记录English

Vision-only UAV State Estimation for Fast Flights Without External Localization Systems

4m 44s688 字数116 segmentsEnglish

完整文本记录

0:01

In this video, we present our approach

0:03

to visiononly UAV state estimation for

0:05

fast and aggressive flights without

0:07

external localization systems. We

0:10

develop a fully onboard estimation

0:12

pipeline using only an IMU and a single

0:14

moninocular camera capable of reliable

0:17

operation during agile flight and GPS

0:19

denied environments. Visual inertial

0:22

odometry or VIO is the standard method

0:25

for onboard state estimation using only

0:27

a camera and an IMU in GPS denied

0:30

environments. However, VIO suffers from

0:33

significant drift and delays during

0:35

aggressive maneuvers. Therefore, we also

0:37

incorporate a landmark detector to

0:39

correct VIO drift using detectable

0:41

landmarks in the environment. At the

0:43

start of the flight, VIO is initialized

0:46

at the UAV's position and defines its

0:49

own coordinate frame, which is connected

0:51

to the world frame through a static

0:53

transformation. As the UAV begins flying

0:56

and performs fast aggressive maneuvers,

0:58

VIO starts to drift and its estimated

1:01

states diverge from the ground truth

1:03

states across all six degrees of

1:04

freedom. Relying on VIO alone for state

1:07

estimation often leads to crashes.

1:10

Current state-of-the-art methods either

1:13

rely on inaccurate VIO estimates such as

1:16

linear and angular velocities or the

1:18

UAV's attitude or require more complex

1:21

hardware including stereo cameras and

1:23

rangefinders.

1:25

In contrast, our approach compensates

1:27

for VIO drift across all UAV states

1:30

while using only an RGB camera and an

1:32

IMU. Here is our estimation pipeline.

1:35

VIO uses IMU and camera data to provide

1:38

drifting UAV states which are fused with

1:41

camera measurements from the landmark

1:42

detector to estimate VIO drift. Then we

1:46

correct the VIO odometry using the

1:48

estimated drift and fuse it with IMU

1:51

data to reduce delay and capture

1:53

aggressive UAV motion. Finally, the

1:56

estimated states are used by the

1:58

controller to track the pre-planned

2:00

trajectory.

2:02

In our paper, we propose a novel model

2:04

of VIO drift, which is incorporated into

2:06

a Calman filter to estimate the drift.

2:09

We then fuse data from VIO, the

2:11

estimated VIO drift, and the IMU to

2:14

produce the final UAV state estimate. As

2:17

you can see in the equations,

2:19

our approach was successfully deployed

2:21

at the A2RL drone racing challenge 2025

2:24

in Abu Dhabi, where we advanced through

2:26

the quarterfinals and semi-finals to

2:28

reach the final round among the top four

2:30

teams out of a total of 210. The goal of

2:34

each round was to complete two laps

2:36

through a predefined sequence of 11

2:38

gates, and we completed multiple twolap

2:40

runs at speeds of up to 45 kmh. Here you

2:43

can see one of our flights. The

2:45

three-dimensional plot in the top left

2:47

corner of this flight shows that the VIO

2:50

estimate shown in gray is insufficient

2:52

for agile flight in cluttered GPS denied

2:55

environments. In contrast, our approach

2:58

provides accurate state estimates shown

3:00

by the blue to red trajectory indicating

3:03

speed from slowest in blue to fastest in

3:05

red. We also performed real world

3:08

experiments on an outdoor track to

3:10

compare our approach against ground

3:12

truth values obtained from RTK. The

3:15

outdoor track consisted of six gates and

3:17

the UAV was required to complete two

3:19

laps. The three-dimensional plot in the

3:22

top left corner shows ground truth data

3:23

from RTK, estimates from VIO and values

3:27

from our approach where color indicates

3:29

speed. Our approach tracks the ground

3:31

truth smoothly while VIO exhibits

3:34

significant drift. We conducted numerous

3:37

flights and performed a statistical

3:39

evaluation comparing our method with

3:41

state-of-the-art approaches and RTK

3:43

values. Here is the table presenting the

3:46

statistical evaluation of our approach

3:48

compared to ground truth values and

3:50

state-of-the-art methods across all UAV

3:52

states including position, orientation,

3:56

linear velocity, and angular velocity.

3:59

Compared to state-of-the-art methods,

4:01

our approach reduces the root mean

4:03

square error of linear velocity by 16%,

4:06

orientation by 70% and angular velocity

4:09

by 88%.

4:11

Our novel approach for visiononly UAV

4:13

state estimation presents an accurate

4:15

onboard pipeline for fast and aggressive

4:17

flights using only a moninocular camera

4:20

and an IMU. Our approach achieves

4:23

significant improvements in linear

4:24

velocity, orientation, and angular

4:27

velocity estimation accuracy in terms of

4:29

root mean square error compared to

4:31

current state-of-the-art methods.

4:33

Additionally, it incorporates a novel

4:35

drift model and directly fuses IMU data

4:39

into the final UAV state estimate.

解锁更多

免费注册以访问高级功能

互动查看器

观看带有同步字幕、可调节叠加层和完整播放控制的视频。

免费注册以解锁

AI 摘要

获取由 AI 立即生成的视频内容摘要、要点和结论。

免费注册以解锁

翻译

一键将字幕翻译成 100 多种语言。以任何格式下载。

免费注册以解锁

思维导图

将字幕可视化为交互式思维导图。一目了然地了解结构。

免费注册以解锁

与字幕聊天

提出关于视频内容的问题。直接从字幕中获取由 AI 驱动的答案。

免费注册以解锁

从您的字幕中获得更多

免费注册并解锁交互式查看器、AI 摘要、翻译、思维导图等。无需信用卡。

    Vision-only UAV State Estimatio… - 完整文字记录 | YouTubeTranscript.dev