Self-supervised Deep LiDAR Odometry for Robotic Applications

Overview

DeLORA: Self-supervised Deep LiDAR Odometry for Robotic Applications

Overview

This is the corresponding code to the above paper ("Self-supervised Learning of LiDAR Odometry for Robotic Applications") which is published at the International Conference on Robotics and Automation (ICRA) 2021. The code is provided by the Robotics Systems Lab at ETH Zurich, Switzerland.

** Authors:** Julian Nubert ([email protected]) , Shehryar Khattak , Marco Hutter

title_img

Copyright IEEE

Python Setup

We provide a conda environment for running our code.

Conda

The conda environment is very comfortable to use in combination with PyTorch because only NVidia drivers are needed. The Installation of suitable CUDA and CUDNN libraries is all handle by Conda.

  • Install conda: link
  • To set up the conda environment run the following command:
conda env create -f conda/DeLORA-py3.9.yml

This installs an environment including GPU-enabled PyTorch, including any needed CUDA and cuDNN dependencies.

  • Activate the environment:
conda activate DeLORA-py3.9
  • Install the package to set all paths correctly:
pip3 install -e .

ROS Setup

For running ROS code in the ./src/ros_utils/ folder you need to have ROS installed (link). We recommend Ubuntu 20.04 and ROS Noetic due to its native Python3 support. For performing inference in Python2.7, convert your PyTorch model with ./scripts/convert_pytorch_models.py and run an older PyTorch version (<1.3).

ros-numpy

In any case you need to install ros-numpy if you want to make use of the provided rosnode:

sudo apt install ros-<distro>-ros-numpy

Datasets and Preprocessing

Instructions on how to use and preprocess the datasets can be found in the ./datasets/ folder. We provide scripts for doing the preprocessing for:

  1. general rosbags containing LiDAR scans,
  2. and for the KITTI dataset in its own format.

Example: KITTI Dataset

LiDAR Scans

Download the "velodyne laster data" from the official KITTI odometry evaluation ( 80GB): link. Put it to <delora_ws>/datasets/kitti, where kitti contains /data_odometry_velodyne/dataset/sequences/00..21.

Groundtruth poses

Please also download the groundtruth poses here. Make sure that the files are located at <delora_ws>/datasets/kitti, where kitti contains /data_odometry_poses/dataset/poses/00..10.txt.

Preprocessing

In the file ./config/deployment_options.yaml make sure to set datasets: ["kitti"]. Then run

preprocess_data.py

Custom Dataset

If you want to add an own dataset please add its sensor specifications to ./config/config_datasets.yaml and ./config/config_datasets_preprocessing.yaml. Information that needs to be added is the dataset name, its sequences and its sensor specifications such as vertical field of view and number of rings.

Deploy

After preprocessing, for each dataset we assume the following hierarchical structure: dataset_name/sequence/scan (see previous dataset example). Our code natively supports training and/or testing on various datasets with various sequences at the same time.

Training

Run the training with the following command:

run_training.py

The training will be executed for the dataset(s) specified in ./config/deployment_options.yaml. You will be prompted to enter a name for this training run, which will be used for reference in the MLFlow logging.

Custom Settings

For custom settings and hyper-parameters please have a look in ./config/.

By default loading from RAM is disabled. If you have enough memory, enable it in ./config/deployment_options.yaml. When loading from disk, the first few iterations are sometimes slow due to I/O, but it should accelerate quite quickly. For storing the KITTI training set entirely in memory, roughly 50GB of RAM are required.

Continuing Training

For continuing training provide the --checkpoint flag with a path to the model checkpoint to the script above.

Visualizing progress and results

For visualizing progress we use MLFlow. It allows for simple logging of parameters, metrics, images, and artifacts. Artifacts could e.g. also be whole TensorBoard logfiles. To visualize the training progress execute (from DeLORA folder):

mlflow ui 

The MLFlow can then be visualized in your browser following the link in the terminal.

Testing

Testing can be run along the line:

run_testing.py --checkpoint <path_to_checkpoint>

The checkpoint can be found in MLFlow after training. It runs testing for the dataset specified in ./config/deployment_options.yaml.

We provide an exemplary trained model in ./checkpoints/kitti_example.pth.

ROS-Node

This ROS-node takes the pretrained model at location <model_location> and performs inference; i.e. it predicts and publishes the relative transformation between incoming point cloud scans. The variable <dataset> should contain the name of the dataset in the config files, e.g. kitti, in order to load the corresponding parameters. Topic and frame names can be specified in the following way:

run_rosnode.py --checkpoint <model_location> --dataset <dataset> --lidar_topic=<name_of_lidar_topic> --lidar_frame=<name_of_lidar_frame>

The resulting odometry will be published as a nav_msgs.msg.Odometry message under the topic /delora/odometry .

Example: DARPA Dataset

For the darpa dataset this could look as follows:

run_rosnode.py --checkpoint ~/Downloads/checkpoint_epoch_0.pth --dataset darpa --lidar_topic "/sherman/lidar_points" --lidar_frame sherman/ouster_link

Comfort Functions

Additional functionalities are provided in ./bin/ and ./scripts/.

Visualization of Normals (mainly for debugging)

Located in ./bin/, see the readme-file ./dataset/README.md for more information.

Creation of Rosbags for KITTI Dataset

After starting a roscore, conversion from KITTI dataset format to a rosbag can be done using the following command:

python scripts/convert_kitti_to_rosbag.py

The point cloud scans will be contained in the topic "/velodyne_points", located in the frame velodyne. E.g. for the created rosbag, our provided rosnode can be run using the following command:

run_rosnode.py --checkpoint ~/Downloads/checkpoint_epoch_30.pth --lidar_topic "/velodyne_points" --lidar_frame "velodyne"

Convert PyTorch Model to older PyTorch Compatibility

Converion of the new model <path_to_model>/model.pth to old (compatible with < PyTorch1.3) <path_to_model>/model_py27.pth can be done with the following:

python scripts/convert_pytorch_models.py --checkpoint <path_to_model>/model

Note that there is no .pth ending in the script.

Time The Network

The execution time of the network can be timed using:

python scripts/time_network.py

Paper

Thank you for citing DeLORA (ICRA-2021) if you use any of this code.

@inproceedings{nubert2021self,
  title={Self-supervised Learning of LiDAR Odometry for Robotic Applications},
  author={Nubert, Julian and Khattak, Shehryar and Hutter, Marco},
  booktitle={IEEE International Conference on Robotics and Automation (ICRA)},
  year={2021},
  organization={IEEE}
}

Dependencies

Dependencies are specified in ./conda/DeLORA-py3.9.yml and ./pip/requirements.txt.

Tuning

If the result does not achieve the desired performance, please have a look at the normal estimation, since the loss is usually dominated by the plane-to-plane loss, which is impacted by noisy normal estimates. For the results presented in the paper we picked some reasonable parameters without further fine-tuning, but we are convinced that less noisy normal estimates would lead to an even better convergence.

Owner
Robotic Systems Lab - Legged Robotics at ETH Zürich
The Robotic Systems Lab investigates the development of machines and their intelligence to operate in rough and challenging environments.
Robotic Systems Lab - Legged Robotics at ETH Zürich
Distance Encoding for GNN Design

Distance-encoding for GNN design This repository is the official PyTorch implementation of the DEGNN and DEAGNN framework reported in the paper: Dista

172 Nov 08, 2022
Implementation of the paper "Language-agnostic representation learning of source code from structure and context".

Code Transformer This is an official PyTorch implementation of the CodeTransformer model proposed in: D. Zügner, T. Kirschstein, M. Catasta, J. Leskov

Daniel Zügner 131 Dec 13, 2022
Training Certifiably Robust Neural Networks with Efficient Local Lipschitz Bounds (Local-Lip)

Training Certifiably Robust Neural Networks with Efficient Local Lipschitz Bounds (Local-Lip) Introduction TL;DR: We propose an efficient and trainabl

17 Dec 01, 2022
PyTorch implementation of our ICCV 2021 paper Intrinsic-Extrinsic Preserved GANs for Unsupervised 3D Pose Transfer.

Unsupervised_IEPGAN This is the PyTorch implementation of our ICCV 2021 paper Intrinsic-Extrinsic Preserved GANs for Unsupervised 3D Pose Transfer. Ha

25 Oct 26, 2022
Tutel MoE: An Optimized Mixture-of-Experts Implementation

Project Tutel Tutel MoE: An Optimized Mixture-of-Experts Implementation. Supported Framework: Pytorch Supported GPUs: CUDA(fp32 + fp16), ROCm(fp32) Ho

Microsoft 344 Dec 29, 2022
Learning Logic Rules for Document-Level Relation Extraction

LogiRE Learning Logic Rules for Document-Level Relation Extraction We propose to introduce logic rules to tackle the challenges of doc-level RE. Equip

41 Dec 26, 2022
Quickly comparing your image classification models with the state-of-the-art models (such as DenseNet, ResNet, ...)

Image Classification Project Killer in PyTorch This repo is designed for those who want to start their experiments two days before the deadline and ki

349 Dec 08, 2022
Tensorflow implementation of soft-attention mechanism for video caption generation.

SA-tensorflow Tensorflow implementation of soft-attention mechanism for video caption generation. An example of soft-attention mechanism. The attentio

Paul Chen 153 Nov 14, 2022
Facial recognition project

Facial recognition project documentation Project introduction This project is developed by linuxu. It is a face model recognition project developed ba

Jefferson 2 Dec 04, 2022
[CVPR 2021] Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective with Transformers

[CVPR 2021] Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective with Transformers

Fudan Zhang Vision Group 897 Jan 05, 2023
DeepVoxels is an object-specific, persistent 3D feature embedding.

DeepVoxels is an object-specific, persistent 3D feature embedding. It is found by globally optimizing over all available 2D observations of

Vincent Sitzmann 196 Dec 25, 2022
Course materials for Fall 2021 "CIS6930 Topics in Computing for Data Science" at New College of Florida

Fall 2021 CIS6930 Topics in Computing for Data Science This repository hosts course materials used for a 13-week course "CIS6930 Topics in Computing f

Yoshi Suhara 101 Nov 30, 2022
Machine Learning Models were applied to predict the mass of the brain based on gender, age ranges, and head size.

Brain Weight in Humans Variations of head sizes and brain weights in humans Kaggle dataset obtained from this link by Anubhab Swain. Image obtained fr

Anne Livia 1 Feb 02, 2022
Codes for NAACL 2021 Paper "Unsupervised Multi-hop Question Answering by Question Generation"

Unsupervised-Multi-hop-QA This repository contains code and models for the paper: Unsupervised Multi-hop Question Answering by Question Generation (NA

Liangming Pan 70 Nov 27, 2022
A library of extension and helper modules for Python's data analysis and machine learning libraries.

Mlxtend (machine learning extensions) is a Python library of useful tools for the day-to-day data science tasks. Sebastian Raschka 2014-2020 Links Doc

Sebastian Raschka 4.2k Jan 02, 2023
DCGAN LSGAN WGAN-GP DRAGAN PyTorch

Recommendation Our GAN based work for facial attribute editing - AttGAN. News 8 April 2019: We re-implement these GANs by Tensorflow 2! The old versio

Zhenliang He 408 Nov 30, 2022
Implementation of Invariant Point Attention, used for coordinate refinement in the structure module of Alphafold2, as a standalone Pytorch module

Invariant Point Attention - Pytorch Implementation of Invariant Point Attention as a standalone module, which was used in the structure module of Alph

Phil Wang 113 Jan 05, 2023
Depth image based mouse cursor visual haptic

Depth image based mouse cursor visual haptic How to run it. Install pyqt5. Install python modules pip install Pillow pip install numpy For illustrati

Xiong Jie 17 Dec 20, 2022
[ICCV 2021] FaPN: Feature-aligned Pyramid Network for Dense Image Prediction

FaPN: Feature-aligned Pyramid Network for Dense Image Prediction [arXiv] [Project Page] @inproceedings{ huang2021fapn, title={{FaPN}: Feature-alig

EMI-Group 175 Dec 30, 2022
PyTorch implementation of MICCAI 2018 paper "Liver Lesion Detection from Weakly-labeled Multi-phase CT Volumes with a Grouped Single Shot MultiBox Detector"

Grouped SSD (GSSD) for liver lesion detection from multi-phase CT Note: the MICCAI 2018 paper only covers the multi-phase lesion detection part of thi

Sang-gil Lee 36 Oct 12, 2022