All Projects → wvangansbeke → Sparse Depth Completion

wvangansbeke / Sparse Depth Completion

Licence: other
Predict dense depth maps from sparse and noisy LiDAR frames guided by RGB images. (Ranked 1st place on KITTI)

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Sparse Depth Completion

direct lidar odometry
Direct LiDAR Odometry: Fast Localization with Dense Point Clouds
Stars: ✭ 202 (-25.74%)
Mutual labels:  lidar
BtcDet
Behind the Curtain: Learning Occluded Shapes for 3D Object Detection
Stars: ✭ 104 (-61.76%)
Mutual labels:  lidar
lidar transfer
Code for Langer et al. "Domain Transfer for Semantic Segmentation of LiDAR Data using Deep Neural Networks", IROS, 2020.
Stars: ✭ 54 (-80.15%)
Mutual labels:  lidar
lidar body tracking
ROS Catkin package to track people using octree and cluster extraction
Stars: ✭ 68 (-75%)
Mutual labels:  lidar
sensor-fusion
Filters: KF, EKF, UKF || Process Models: CV, CTRV || Measurement Models: Radar, Lidar
Stars: ✭ 96 (-64.71%)
Mutual labels:  lidar
Awesome-3D-Object-Detection-for-Autonomous-Driving
Papers on 3D Object Detection for Autonomous Driving
Stars: ✭ 52 (-80.88%)
Mutual labels:  lidar
point-cloud-prediction
Self-supervised Point Cloud Prediction Using 3D Spatio-temporal Convolutional Networks
Stars: ✭ 97 (-64.34%)
Mutual labels:  lidar
3dfier
The open-source tool for creating of 3D models
Stars: ✭ 260 (-4.41%)
Mutual labels:  lidar
awesome-lidar
😎 Awesome LIDAR list. The list includes LIDAR manufacturers, datasets, point cloud-processing algorithms, point cloud frameworks and simulators.
Stars: ✭ 217 (-20.22%)
Mutual labels:  lidar
Pandora SDK
Development kit for Pandora
Stars: ✭ 14 (-94.85%)
Mutual labels:  lidar
Applications-LidarBot
No description or website provided.
Stars: ✭ 15 (-94.49%)
Mutual labels:  lidar
urban road filter
Real-time LIDAR-based Urban Road and Sidewalk detection for Autonomous Vehicles 🚗
Stars: ✭ 134 (-50.74%)
Mutual labels:  lidar
rlas
R package to read and write las and laz files used to store LiDAR data
Stars: ✭ 23 (-91.54%)
Mutual labels:  lidar
camera lidar calibration
A tool used for calibrate the extrinsic between 2D laser range finder (LRF) and camera. ROS Version: https://github.com/TurtleZhong/camera_lidar_calibration_v2
Stars: ✭ 48 (-82.35%)
Mutual labels:  lidar
Pandar40 SDK
Development kit for Pandar40
Stars: ✭ 20 (-92.65%)
Mutual labels:  lidar
ple
Probabilistic line extraction from 2-D range scan
Stars: ✭ 47 (-82.72%)
Mutual labels:  lidar
UrbanLoco
UrbanLoco: A Full Sensor Suite Dataset for Mapping and Localization in Urban Scenes
Stars: ✭ 147 (-45.96%)
Mutual labels:  lidar
3d cnn tensorflow
KITTI data processing and 3D CNN for Vehicle Detection
Stars: ✭ 266 (-2.21%)
Mutual labels:  lidar
Veloview
VeloView performs real-time visualization and easy processing of live captured 3D LiDAR data from Velodyne sensors (Alpha Prime™, Puck™, Ultra Puck™, Puck Hi-Res™, Alpha Puck™, Puck LITE™, HDL-32, HDL-64E). Runs on Windows, Linux and MacOS
Stars: ✭ 253 (-6.99%)
Mutual labels:  lidar
tloam
T-LOAM: Truncated Least Squares Lidar-only Odometry and Mapping in Real-Time
Stars: ✭ 164 (-39.71%)
Mutual labels:  lidar

Sparse-Depth-Completion

This repo contains the implementation of our paper Sparse and Noisy LiDAR Completion with RGB Guidance and Uncertainty by Wouter Van Gansbeke, Davy Neven, Bert De Brabandere and Luc Van Gool.

If you find this interesting or relevant to your work, consider citing:

@inproceedings{wvangansbeke_depth_2019,
    author={Van Gansbeke, Wouter and Neven, Davy and De Brabandere, Bert and Van Gool, Luc},
    booktitle={2019 16th International Conference on Machine Vision Applications (MVA)},
    title={Sparse and Noisy LiDAR Completion with RGB Guidance and Uncertainty},
    year={2019},
    pages={1-6},
    organization={IEEE}
}

License

This software is released under a creative commons license which allows for personal and research use only. For a commercial license please contact the authors. You can view a license summary here

Introduction

Monocular depth prediction methods fail to generate absolute and precise depth maps and stereoscopic approaches are still significantly outperformed by LiDAR based approaches. The goal of the depth completion task is to generate dense depth predictions from sparse and irregular point clouds. This project makes use of uncertainty to combine multiple sensor data in order to generate accurate depth predictions. Mapped lidar points together with RGB images (monococular) are used in this framework. This method holds the 1st place entry on the KITTI depth completion benchmark at the time of submission of the paper.

The contribution of this paper is threefold:

  • Global and local information are combined in order to accurately complete and correct the sparse and noisy LiDAR input. Monocular RGB images are used for the guidance of this depth completion task.
  • Confidence maps are learned for the global branch and the local branch in an unsupervised manner. The predicted depth maps are weighted by their respective confidence map. This is the late fusion technique used in our framework.
  • This method ranks first on the KITTI depth completion benchmark without using additional data or postprocessing.

See full demo on YouTube. The predictions of our model for the KITTI test set can be downloaded here.

demo

Requirements

Python 3.7 The most important packages are pytorch, torchvision, numpy, pillow and matplotlib. (Works with Pytorch 1.1)

Dataset

The Kitti dataset has been used. First download the dataset of the depth completion. Secondly, you'll need to unzip and download the camera images from kitti. I used the file download_raw_files.sh, but this is at your own risk. Make sure you understand it, otherwise don't use it. If you want to keep it safe, go to kitti's website.

The complete dataset consists of 85898 training samples, 6852 validation samples, 1000 selected validation samples and 1000 test samples.

Preprocessing

This step is optional, but allows you to transform the images to jpgs and to downsample the original lidar frames. This will create a new dataset in $dest. You can find the required preprocessing in: Datasets/Kitti_loader.py

Run:

source Shell/preprocess $datapath $dest $num_samples

(Firstly, I transformed the png's to jpg - images to save place. Secondly, two directories are built i.e. one for training and one for validation. See Datasets/Kitti_loader.py)

Dataset structure should look like this:

|--depth selection
|-- Depth
     |-- train
           |--date
               |--sequence1
               | ...
     |--validation
|--RGB
    |--train
         |--date
             |--sequence1
             | ...
    |--validation

Run Code

To run the code:

python main.py --data_path /path/to/data/ --lr_policy plateau

Flags:

  • Set flag "input_type" to rgb or depth.
  • Set flag "pretrained" to true or false to use a model pretrained on Cityscapes for the global branch.
  • See python main.py --help for more information.

or

source Shell/train.sh $datapath

checkout more details in the bash file.

Trained models

Our network architecture is based on ERFNet.

You can find the model pretrained on Cityscapes here. This model is used for the global network.

You can find a fully trained model and its corresponding predictions for the KITTI test set here. The RMSE is around 802 mm on the selected validation set for this model as reported in the paper.

To test it: Save the model in a folder in the Saved directory.

and execute the following command:

source Test/test.sh /path/to/directory_with_saved_model/ $num_samples /path/to/dataset/ /path/to/directory_with_ground_truth_for_selected_validation_files/

(You might have to recompile the C files for testing, provided by KITTI, if your architecture is different from mine)

Results

Comparision with state-of-the-art:

results

Discussion

Practical discussion:

  • I recently increased the stability of the training process and I also made the convergence faster by adding some skip connections between the global and local network. Initially I only used guidance by multiplication with an attention map (=probability), but found out that it is less robust and that differences between a focal MSE and vanilla MSE loss function were now negligible. Be aware that this change will alter the appearance of the confidence maps since fusion happens at mutliple stages now.

  • Feel free to experiment with different architectures for the global or local network. It is easy to add new architectures to Models/__init__.py

  • I used a Tesla V100 GPU for evaluation.

Acknowledgement

This work was supported by Toyota, and was carried out at the TRACE Lab at KU Leuven (Toyota Research on Automated Cars in Europe - Leuven)

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].