Abstract
We present COIN-LIO, a LiDAR Inertial Odometry pipeline that tightly couples information from LiDAR intensity with geometry-based point cloud registration. The focus of our work is to improve the robustness of LiDAR-inertial odometry in geometrically degenerate scenarios, like tunnels or flat fields. We project LiDAR intensity returns into an intensity image, and propose an image processing pipeline that produces filtered images with improved brightness consistency within the image as well as across different scenes. To effectively leverage intensity as an additional modality, we present a novel feature selection scheme that detects uninformative directions in the point cloud registration and explicitly selects patches with complementary image information. Photometric error minimization in the image patches is then fused with inertial measurements and point-to-plane registration in an iterated Extended Kalman Filter. The proposed approach improves accuracy and robustness on a public dataset. We additionally publish a new dataset, that captures five real-world environments in challenging, geometrically degenerate scenes. By using the additional photometric information, our approach shows drastically improved robustness against geometric degeneracy in environments where all compared baseline approaches fail.
Please cite our work if you are using COIN-LIO in your research.
@inproceedings{pfreundschuh2024coin,
title={COIN-LIO: Complementary Intensity-Augmented LiDAR Inertial Odometry},
author={Pfreundschuh, Patrick and Oleynikova, Helen and Cadena, Cesar and Siegwart, Roland and Andersson, Olov},
booktitle={2024 IEEE International Conference on Robotics and Automation (ICRA)},
pages={1730--1737},
year={2024},
organization={IEEE}
}
This package was developed on Ubuntu 20.04 using ROS Noetic. Other versions should also work but have not been tested and we do not guarantee support.
- If not done yet, please install ROS, install the proposed system dependencies.
Install some additional system dependencies:
sudo apt-get install python3-catkin-tools libgoogle-glog-dev
- Then create a catkin workspace:
mkdir -p ~/catkin_ws/src cd ~/catkin_ws catkin init catkin config --extend /opt/ros/$ROS_DISTRO catkin config --cmake-args -DCMAKE_BUILD_TYPE=RelWithDebInfo catkin config --merge-devel
- Clone COIN-LIO into your workspace:
cd ~/catkin_ws/src git clone [email protected]:ethz-asl/coin-lio.git cd COIN-LIO
- Build COIN-LIO:
catkin build coin_lio
To instead use docker, check out the repository locally, navigate to it, and:
bash cd docker/ ./run_docker.sh -b
Which will build a docker image with a copy of the code checked out inside.
Your ~/data
folder will be mounted to /root/data
within the docker, so you
can download datasets and follow the rest of the tutorial below. On future runs,
you can simply use ./run_docker.sh
(without -b
) to not re-build the image.
The ENWIDE dataset sequences can be downloaded here. Run a sequence:
roslaunch coin_lio mapping_enwide.launch bag_file:=<example_bag_path.bag>
The Newer College Dataset sequences can be downloaded here. Run a sequence:
roslaunch coin_lio mapping_newer_college.launch bag_file:=<example_bag_path.bag>
Note on LiDAR type: COIN-LIO currently only supports data from Ouster LiDARs, as we use the calibration in the metadata file for the image projection model. Implementing different sensors is theoretically possible but requires a proper implementation of a projection model that works for the specific sensor. Contributions are welcome.
- LiDAR: Since different Ouster sensors have different image projection parameters, we need to run a calibration tool to evaluate the column shift which is required to correct the image projection model. This procedure is only required once per sensor.
It is important to use the metadata file that corresponds to your specific sensor (more information can be found here).
roslaunch coin_lio calibrate.launch bag_file:=<bag_path.bag> metadata_file:=<metadata_path.json> point_topic:=<pointcloud_topic>
The evaluated column shift parameter will be printed at the end of the procedure.
- IMU: If you are not using the built-in IMU in the Ouster LiDAR, you need to adapt the extrinsic calibration between IMU and LiDAR accordingly in the parameter file.
Launch with settings for your data:
roslaunch coin_lio mapping.launch metadata_file:=<metadata_path.json> column_shift:=<parameter from calibration> point_topic:=<pointcloud_topic> imu_topic:=<imu_topic>
Play your data:
rosbag play <bag_path.bag>
The line artifact removal filter can be tested and tuned using the provided notebook:
COIN-LIO builds on top of FAST-LIO2 for the point-to-plane registration. Our dashboard was inspired by DLIO. We thank the authors for open-sourcing their outstanding works.
We used ascii-image-converter for our ascii animation.