All Projects → moabitcoin → Holy Edge

moabitcoin / Holy Edge

Licence: gpl-3.0
Holistically-Nested Edge Detection

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Holy Edge

Awesome Carla
👉 CARLA resources such as tutorial, blog, code and etc https://github.com/carla-simulator/carla
Stars: ✭ 246 (-11.19%)
Mutual labels:  deeplearning, segmentation, detection
Cnn Paper2
🎨 🎨 深度学习 卷积神经网络教程 :图像识别,目标检测,语义分割,实例分割,人脸识别,神经风格转换,GAN等🎨🎨 https://dataxujing.github.io/CNN-paper2/
Stars: ✭ 77 (-72.2%)
Mutual labels:  deeplearning, segmentation, detection
Awesome Gan For Medical Imaging
Awesome GAN for Medical Imaging
Stars: ✭ 1,814 (+554.87%)
Mutual labels:  deeplearning, segmentation, detection
Jacinto Ai Devkit
Training & Quantization of embedded friendly Deep Learning / Machine Learning / Computer Vision models
Stars: ✭ 49 (-82.31%)
Mutual labels:  deeplearning, segmentation, detection
Paddlex
PaddlePaddle End-to-End Development Toolkit(『飞桨』深度学习全流程开发工具)
Stars: ✭ 3,399 (+1127.08%)
Mutual labels:  deeplearning, segmentation, detection
image segmentation dl
🍞 基于深度学习方法的图像分割(含语义分割、实例分割、全景分割)。
Stars: ✭ 76 (-72.56%)
Mutual labels:  segmentation, deeplearning
rgbd person tracking
R-GBD Person Tracking is a ROS framework for detecting and tracking people from a mobile robot.
Stars: ✭ 46 (-83.39%)
Mutual labels:  detection, segmentation
Awesome-Vision-Transformer-Collection
Variants of Vision Transformer and its downstream tasks
Stars: ✭ 124 (-55.23%)
Mutual labels:  detection, segmentation
unsupervised llamas
Code for https://unsupervised-llamas.com
Stars: ✭ 70 (-74.73%)
Mutual labels:  detection, segmentation
mmrazor
OpenMMLab Model Compression Toolbox and Benchmark.
Stars: ✭ 644 (+132.49%)
Mutual labels:  detection, segmentation
mri-deep-learning-tools
Resurces for MRI images processing and deep learning in 3D
Stars: ✭ 56 (-79.78%)
Mutual labels:  detection, segmentation
crowd density segmentation
The code for preparing the training data for crowd counting / segmentation algorithm.
Stars: ✭ 21 (-92.42%)
Mutual labels:  detection, segmentation
BCNet
Deep Occlusion-Aware Instance Segmentation with Overlapping BiLayers [CVPR 2021]
Stars: ✭ 434 (+56.68%)
Mutual labels:  detection, segmentation
rembg-greenscreen
Rembg Video Virtual Green Screen Edition
Stars: ✭ 210 (-24.19%)
Mutual labels:  segmentation, deeplearning
Shadowless
A Fast and Open Source Autonomous Perception System.
Stars: ✭ 29 (-89.53%)
Mutual labels:  detection, segmentation
volkscv
A Python toolbox for computer vision research and project
Stars: ✭ 58 (-79.06%)
Mutual labels:  detection, segmentation
dilation-keras
Multi-Scale Context Aggregation by Dilated Convolutions in Keras.
Stars: ✭ 72 (-74.01%)
Mutual labels:  segmentation, deeplearning
Deep-Learning
It contains the coursework and the practice I have done while learning Deep Learning.🚀 👨‍💻💥 🚩🌈
Stars: ✭ 21 (-92.42%)
Mutual labels:  deeplearning, vgg16
Yolov5-deepsort-driverDistracted-driving-behavior-detection
基于深度学习的驾驶员分心驾驶行为(疲劳+危险行为)预警系统使用YOLOv5+Deepsort实现驾驶员的危险驾驶行为的预警监测
Stars: ✭ 107 (-61.37%)
Mutual labels:  detection, deeplearning
Jejunet
Real-Time Video Segmentation on Mobile Devices with DeepLab V3+, MobileNet V2. Worked on the project in 🏝 Jeju island
Stars: ✭ 258 (-6.86%)
Mutual labels:  deeplearning, segmentation

Holistically-Nested Edge Detection

The code is also hosted at https://github.com/harsimrat-eyeem/holy-edge - Harsimrat Sandhawalia

Input image Final fused Edge maps Edge maps from side layers
  • Final fused edge maps generated at 100, 500, 1k, 1.5k, 2k, 2.5k, 3k, 3.5k, 4k, 4.5k iterations.
  • Edge maps form side layers generated at 5k iterations.

This repository contains tensorflow implementation of the HED model. Details of hyper-paramters are available in the paper

@InProceedings{xie15hed,
  author = {"Xie, Saining and Tu, Zhuowen"},
  Title = {Holistically-Nested Edge Detection},
  Booktitle = "Proceedings of IEEE International Conference on Computer Vision",
  Year  = {2015},
}

Get this repo

git clone https://github.com/harsimrat-eyeem/holy-edge.git

Installing requirements

Its recommended to install the requirements in a conda virtual environment

cd holy-edge
pip install -r requirements.txt
export OMP_NUM_THREADS=1

Setting up

Edit the config file located at hed/configs/hed.yaml. Set the paths below. Make sure the directories exist and you have read/write permissions on them. The HED model is trained on augmented training set created by the authors.

# location where training data : http://vcl.ucsd.edu/hed/HED-BSDS.tar would be downloaded and decompressed
download_path: '<path>'
# location of snapshot and tensorbaord summary events
save_dir: '<path>'
# location where to put the generated edgemaps during testing
test_output: '<path>'

Training data & Models

You can train the model or you can fetch a pre-trained HED snapshot provided in the repo to simply generate edgemaps.

Fetch VGG-16 models weights trained on ImageNet via git-lfs

git lfs fetch && git lfs pull
md5sum hed/models/vgg16.npy

19ff134af12b6ea0c0ff35664b031ba5 hed/models/vgg16.npy

This downloads the augmented training set created by authors of HED. Augmentation strategies include rotation to 16 predefined angles and cropping largest rectangle from the image. Details in section (4.1). To download training data run

python run-hed.py --download-data --config-file hed/configs/hed.yaml

VGG-16 base model

VGG base model available here is used for producing multi-level features. The model is modified according with Section (3.) of the paper. Deconvolution layers are set with tf.nn.conv2d_transpose. The model uses single deconvolution layer in each side layers. Another implementation uses stacked bilinear deconvolution layers. In this implementation the upsampling parameters are learned while finetuning of the model. Pre-trained weights for VGG-16 are hosted with git-lfs in this repo.

Training

If you want to run predictions on pre-trained HED model, skip this and go to Testing with pre-trained model

Launch training

CUDA_VISIBLE_DEVICES=0 python run-hed.py --train --config-file hed/configs/hed.yaml

Launch tensorboard

tensorboard --logdir=<save_dir>

Testing

Edit the snapshot you want to use for testing in hed/configs/hed.yaml

test_snapshot: <snapshot number>
CUDA_VISIBLE_DEVICES=1 python run-hed.py --test --config-file hed/configs/hed.yaml --gpu-limit 0.4
feh <test_output>

Testing with pre-trained model

Edit your config file located at hed/configs/hed.yaml to change the location of the pre-trained HED model

save_dir: <path_to_repo_on_disk>/hed
test_snapshot: 5000
# location where to put the generated edgemaps during testing
test_output: '<path>'

Run predictions

CUDA_VISIBLE_DEVICES=1 python run-hed.py --test --config-file hed/configs/hed.yaml --gpu-limit 0.4
feh <test_output>
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].