Publication:
Loosely coupled 4D-Radar-Inertial Odometry for Ground Robots

Loading...
Thumbnail Image

Publication date

Reading date

Event date

Start date of the public exhibition period

End date of the public exhibition period

Advisors

Authors of photography

Person who provides the photography

Journal Title

Journal ISSN

Volume Title

Publisher

Springer Nature
Export

Research Projects

Organizational Units

Journal Issue

Abstract

Accurate robot odometry is essential for autonomous navigation. While numerous techniques have been developed based on various sensor suites, odometry estimation using only radar and IMU remains an underexplored area. Radar proves particularly valuable in environments where traditional sensors, like cameras or LiDAR, may struggle, especially in low-light conditions or when faced with environmental challenges like fog, rain or smoke. However, despite its robustness, radar data is noisier and more prone to outliers, requiring specialized processing approaches. In this paper, we propose a graph-based optimization approach (https://github.com/robotics-upo/4D-Radar-Odom.git) using a sliding window for radar-based odometry, designed to maintain robust relationships between poses by forming a network of connections, while keeping computational costs fixed (specially beneficial in long trajectories). Additionally, we introduce an enhancement in the ego-velocity estimation specifically for ground vehicles, both holonomic and non-holonomic, which subsequently improves the direct odometry input required by the optimizer. Finally, we present a comparative study of our approach against existing algorithms, showing how our pure odometry approach improves the state of art in all trajectories of the NTU4DRadLM dataset, achieving promising results when evaluating key performance metrics.

Doctoral program

Related publication

Research projects

Description

Proyectos de investigación PID2021-127648OB-C31 TED2021- 132476B-I00

Bibliographic reference

J Intell Robot Syst 111, 107 (2025)

Photography rights