All Projects → intel-isl → Midas

intel-isl / Midas

Licence: mit
Code for robust monocular depth estimation described in "Ranftl et. al., Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer, TPAMI 2020"

Programming Languages

java
68154 projects - #9 most used programming language

Projects that are alternatives of or similar to Midas

Nlp Paper
NLP Paper
Stars: ✭ 484 (-31.73%)
Mutual labels:  deeplearning
Tr
Free Offline OCR 离线的中文文本检测+识别SDK
Stars: ✭ 598 (-15.66%)
Mutual labels:  deeplearning
Ffdl
Fabric for Deep Learning (FfDL, pronounced fiddle) is a Deep Learning Platform offering TensorFlow, Caffe, PyTorch etc. as a Service on Kubernetes
Stars: ✭ 640 (-9.73%)
Mutual labels:  deeplearning
Openvino Yolov3
YoloV3/tiny-YoloV3+RaspberryPi3/Ubuntu LaptopPC+NCS/NCS2+USB Camera+Python+OpenVINO
Stars: ✭ 500 (-29.48%)
Mutual labels:  deeplearning
Deberta
The implementation of DeBERTa
Stars: ✭ 541 (-23.7%)
Mutual labels:  deeplearning
Deeplearning Assignment
深度学习笔记
Stars: ✭ 619 (-12.69%)
Mutual labels:  deeplearning
East
This is a pytorch re-implementation of EAST: An Efficient and Accurate Scene Text Detector.
Stars: ✭ 478 (-32.58%)
Mutual labels:  deeplearning
Deepfacelab
DeepFaceLab is the leading software for creating deepfakes.
Stars: ✭ 30,308 (+4174.75%)
Mutual labels:  deeplearning
Reversi Alpha Zero
Reversi reinforcement learning by AlphaGo Zero methods.
Stars: ✭ 598 (-15.66%)
Mutual labels:  deeplearning
Seq2seq
A general-purpose encoder-decoder framework for Tensorflow
Stars: ✭ 5,455 (+669.39%)
Mutual labels:  deeplearning
Convcrf
This repository contains the reference implementation for our proposed Convolutional CRFs.
Stars: ✭ 514 (-27.5%)
Mutual labels:  deeplearning
Deeplearning
深度学习入门教程, 优秀文章, Deep Learning Tutorial
Stars: ✭ 6,783 (+856.7%)
Mutual labels:  deeplearning
Recommendersystem Paper
This repository includes some papers that I have read or which I think may be very interesting.
Stars: ✭ 619 (-12.69%)
Mutual labels:  deeplearning
Learn Data Science For Free
This repositary is a combination of different resources lying scattered all over the internet. The reason for making such an repositary is to combine all the valuable resources in a sequential manner, so that it helps every beginners who are in a search of free and structured learning resource for Data Science. For Constant Updates Follow me in …
Stars: ✭ 4,757 (+570.94%)
Mutual labels:  deeplearning
Neupy
NeuPy is a Tensorflow based python library for prototyping and building neural networks
Stars: ✭ 670 (-5.5%)
Mutual labels:  deeplearning
Monk v1
Monk is a low code Deep Learning tool and a unified wrapper for Computer Vision.
Stars: ✭ 480 (-32.3%)
Mutual labels:  deeplearning
Deep Image Matting
This is tensorflow implementation for paper "Deep Image Matting"
Stars: ✭ 614 (-13.4%)
Mutual labels:  deeplearning
Ai Series
📚 [.md & .ipynb] Series of Artificial Intelligence & Deep Learning, including Mathematics Fundamentals, Python Practices, NLP Application, etc. 💫 人工智能与深度学习实战,数理统计篇 | 机器学习篇 | 深度学习篇 | 自然语言处理篇 | 工具实践 Scikit & Tensoflow & PyTorch 篇 | 行业应用 & 课程笔记
Stars: ✭ 702 (-0.99%)
Mutual labels:  deeplearning
Deeplearningmugenknock
でぃーぷらーにんぐを無限にやってディープラーニングでDeepLearningするための実装CheatSheet
Stars: ✭ 684 (-3.53%)
Mutual labels:  deeplearning
Deeplearning.ai
deeplearning.ai , By Andrew Ng, All video link
Stars: ✭ 625 (-11.85%)
Mutual labels:  deeplearning

Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer

This repository contains code to compute depth from a single image. It accompanies our paper:

Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer
René Ranftl, Katrin Lasinger, David Hafner, Konrad Schindler, Vladlen Koltun

MiDaS v2.1 was trained on 10 datasets (ReDWeb, DIML, Movies, MegaDepth, WSVD, TartanAir, HRWSI, ApolloScape, BlendedMVS, IRS) with multi-objective optimization. The original model that was trained on 5 datasets (MIX 5 in the paper) can be found here.

Changelog

  • [Nov 2020] Released MiDaS v2.1:
  • [Jul 2020] Added TensorFlow and ONNX code. Added online demo.
  • [Dec 2019] Released new version of MiDaS - the new model is significantly more accurate and robust
  • [Jul 2019] Initial release of MiDaS (Link)

Online demo

An online demo of the model is available: http://35.202.76.57/

Please be patient. Inference might take up to 30 seconds due to hardware restrictions.

Setup

  1. Download the model weights model-f6b98070.pt and model-small-70d6b9c8.pt and place the file in the root folder.

  2. Set up dependencies:

    conda install pytorch torchvision opencv
    

    The code was tested with Python 3.7, PyTorch 1.7.0, and OpenCV 4.4.0.

Usage

  1. Place one or more input images in the folder input.

  2. Run the model:

    python run.py
    

    Or run the small model:

    python run.py --model_weights model-small-70d6b9c8.pt --model_type small
    
  3. The resulting inverse depth maps are written to the output folder.

via Docker

  1. Make sure you have installed Docker and the NVIDIA Docker runtime.

  2. Build the Docker image:

    docker build -t midas .
    
  3. Run inference:

    docker run --rm --gpus all -v $PWD/input:/opt/MiDaS/input -v $PWD/output:/opt/MiDaS/output midas
    

    This command passes through all of your NVIDIA GPUs to the container, mounts the input and output directories and then runs the inference.

via PyTorch Hub

The pretrained model is also available on PyTorch Hub

via TensorFlow or ONNX

See README in the tf subdirectory.

via Mobile (iOS / Android)

See README in the mobile subdirectory.

via ROS1 (Robot Operating System)

See README in the ros subdirectory.

Accuracy

Zero-shot error (the lower - the better) and speed (FPS):

Model DIW, WHDR Eth3d, AbsRel Sintel, AbsRel Kitti, δ>1.25 NyuDepthV2, δ>1.25 TUM, δ>1.25 Speed, FPS
Small models: iPhone 11
MiDaS v2 small 0.1248 0.1550 0.3300 21.81 15.73 17.00 0.6
MiDaS v2.1 small URL 0.1344 0.1344 0.3370 29.27 13.43 14.53 30
Relative improvement -7.7% +13.3% -2.1% -34.2% +14.6% +14.5% 50x
Big models: GPU RTX 2080Ti
MiDaS v2 large URL 0.1246 0.1290 0.3270 23.90 9.55 14.29 59
MiDaS v2.1 large URL 0.1295 0.1155 0.3285 16.08 8.71 12.51 59
Relative improvement -3.9% +10.5% -0.52% +32.7% +8.8% +12.5% 1x

Citation

Please cite our paper if you use this code or any of the models:

@article{Ranftl2020,
	author    = {Ren\'{e} Ranftl and Katrin Lasinger and David Hafner and Konrad Schindler and Vladlen Koltun},
	title     = {Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer},
	journal   = {IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI)},
	year      = {2020},
}

License

MIT License

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].