All Projects → codeslake → DMENet

codeslake / DMENet

Licence: AGPL-3.0 license
[CVPR 2019] Official TensorFlow Implementation for "Deep Defocus Map Estimation using Domain Adaptation"

Programming Languages

python
139335 projects - #7 most used programming language
matlab
3953 projects
shell
77523 projects

Projects that are alternatives of or similar to DMENet

multi-task-defocus-deblurring-dual-pixel-nimat
Reference github repository for the paper "Improving Single-Image Defocus Deblurring: How Dual-Pixel Images Help Through Multi-Task Learning". We propose a single-image deblurring network that incorporates the two sub-aperture views into a multitask framework. Specifically, we show that jointly learning to predict the two DP views from a single …
Stars: ✭ 29 (-66.28%)
Mutual labels:  defocus-blur, defocus-deblurring
recurrent-defocus-deblurring-synth-dual-pixel
Reference github repository for the paper "Learning to Reduce Defocus Blur by Realistically Modeling Dual-Pixel Data". We propose a procedure to generate realistic DP data synthetically. Our synthesis approach mimics the optical image formation found on DP sensors and can be applied to virtual scenes rendered with standard computer software. Lev…
Stars: ✭ 30 (-65.12%)
Mutual labels:  defocus-blur, defocus-deblurring
HierarchyLayout
[CVPR'19] Hierarchy Denoising Recursive Autoencoders for 3D Scene Layout Prediction
Stars: ✭ 20 (-76.74%)
Mutual labels:  cvpr2019
2019-CVPR-AIC-Track-1-UWIPL
Repository for 2019 CVPR AI City Challenge Track 1 from IPL@UW
Stars: ✭ 19 (-77.91%)
Mutual labels:  cvpr2019
Craft Pytorch
Official implementation of Character Region Awareness for Text Detection (CRAFT)
Stars: ✭ 2,220 (+2481.4%)
Mutual labels:  cvpr2019
Learning2AdaptForStereo
Code for: "Learning To Adapt For Stereo" accepted at CVPR2019
Stars: ✭ 73 (-15.12%)
Mutual labels:  cvpr2019
TA3N
[ICCV 2019 Oral] TA3N: https://github.com/cmhungsteve/TA3N (Most updated repo)
Stars: ✭ 45 (-47.67%)
Mutual labels:  cvpr2019
object centric VAD
An Tensorflow Re-Implement of CVPR 2019 "Object-centric Auto-Encoders and Dummy Anomalies for Abnormal Event Detection in Video"
Stars: ✭ 89 (+3.49%)
Mutual labels:  cvpr2019
SKNet-PyTorch
Nearly Perfect & Easily Understandable PyTorch Implementation of SKNet
Stars: ✭ 62 (-27.91%)
Mutual labels:  cvpr2019
Cvpr2021 Paper Code Interpretation
cvpr2021/cvpr2020/cvpr2019/cvpr2018/cvpr2017 论文/代码/解读/直播合集,极市团队整理
Stars: ✭ 8,075 (+9289.53%)
Mutual labels:  cvpr2019
obman
[cvpr19] Hands+Objects synthetic dataset, instructions to download and code to load the dataset
Stars: ✭ 120 (+39.53%)
Mutual labels:  cvpr2019
Siammask
[CVPR2019] Fast Online Object Tracking and Segmentation: A Unifying Approach
Stars: ✭ 3,205 (+3626.74%)
Mutual labels:  cvpr2019
video-clip-order-prediction
Self-supervised Spatiotemporal Learning via Video Clip Order Prediction
Stars: ✭ 85 (-1.16%)
Mutual labels:  cvpr2019
CompenNet
[CVPR'19] End-to-end Projector Photometric Compensation
Stars: ✭ 35 (-59.3%)
Mutual labels:  cvpr2019
understanding apr
No description or website provided.
Stars: ✭ 38 (-55.81%)
Mutual labels:  cvpr2019
unsup temp embed
Unsupervised learning of action classes with continuous temporal embedding (CVPR'19)
Stars: ✭ 62 (-27.91%)
Mutual labels:  cvpr2019
video repres mas
code for CVPR-2019 paper: Self-supervised Spatio-temporal Representation Learning for Videos by Predicting Motion and Appearance Statistics
Stars: ✭ 63 (-26.74%)
Mutual labels:  cvpr2019
PSENet-Tensorflow
TensorFlow implementation of PSENet text detector (Shape Robust Text Detection with Progressive Scale Expansion Networkt)
Stars: ✭ 51 (-40.7%)
Mutual labels:  cvpr2019
art2real
Art2Real: Unfolding the Reality of Artworks via Semantically-Aware Image-to-Image Translation. CVPR 2019
Stars: ✭ 71 (-17.44%)
Mutual labels:  cvpr2019
visual-compatibility
Context-Aware Visual Compatibility Prediction (https://arxiv.org/abs/1902.03646)
Stars: ✭ 92 (+6.98%)
Mutual labels:  cvpr2019

DMENet: Deep Defocus Map Estimation Network
Official Implementation of the CVPR 2021 Paper
Project | Paper | Supp | Poster
License CC BY-NC

This repository contains the official matlab implementation of SYNDOF generation used in the following paper:

Deep Defocus Map Estimation using Domain Adaptation
Junyong Lee1, Sungkil Lee2, Sunghyun Cho3, and Seungyong Lee1
1POSTECH, 2Sungkyunkwan University, 3DGIST
IEEE Computer Vision and Pattern Recognition (CVPR) 2019

Getting Started

Prerequisites

Tested environment

Ubuntu Python 3.6 TensorFlow 1.15.0 TensorLayer 1.11.1 CUDA 10.0.130 CUDNN 7.6.

  1. Setup environment

    $ git clone https://github.com/codeslake/DMENet.git
    $ cd DMENet
    
    # for CUDA10.0
    $ conda create -y --name DMENet python=3.6 && conda activate DMENet
    $ sh install_CUDA10.0.sh
    
    # for CUDA11.1 (the name of conda environment matters)
    $ conda create -y --name DMENet_CUDA11 python=3.6 && conda activate DMENet_CUDA11
    $ sh install_CUDA11.1.sh
  2. Download and unzip datasets(Google Drive | Dropbox | OneDrive) under [DATASET_ROOT].

    [DATASET_ROOT]
     ├── train
     │   ├── SYNDOF
     │   ├── CUHK
     │   └── Flickr
     └── test
         ├── CUHK
         ├── RTF
         └── SYNDOF
    

    Note:

    • [DATASET_ROOT] is currently set to ./datasets/. It can be specified by modifying config.data_offset in ./config.py.
  3. Download pretrained weights of DMENet (Google Drive | Dropbox) and unzip it as in [LOG_ROOT]/DMENet_BDCS/checkpoint/DMENet_BDCS.npz ([LOG_ROOT] is currently set to ./logs/).

  4. Download pretrained VGG19 weights (Google Drive | Dropbox) and unzip as in pretrained/vgg19.npy (for training only).

Logs

  • Training and testing logs will be saved under [LOG_ROOT]/[mode]/:

    [LOG_ROOT]
     ├── [mode]
     │   ├── checkpoint      # model checkpoint
     │   ├── log             # scalar/image log for tensorboard
     │   ├── sample          # sample images of training
     │   └── result          # resulting images of evaluation
     └── ...
    

    [LOG_ROOT] can be modified with config.root_offset in ./config.py.

Testing final model of CVPR 2019

Please note that due to the server issue, the checkpoint used for the paper is lost.
The provided checkpoint is the new checkpoint that shows the closest evaluation results as in the paper.

Check out updated performance with the new checkpoint.

  • Test the final model by:

    python main.py --mode DMENet_BDCS --test_set CUHK

    Testing results will be saved in [LOG_ROOT]/[mode]/result/[test_set]/:

    ...
    [LOG_ROOT]/[mode]/result/
     └── [test_set]
         ├── image                     # input defocused images
         ├── defocus_map               # defocus images (network's direct output in range [0, 1])
         ├── defocus_map_min_max_norm  # min-max normalized defocus images in range [0, 1] for visualization
         └── sigma_map_7_norm          # sigma maps containing normalized standard deviations (in range [0, 1]) for a Gaussian kernel. For the actual standard deviation value, one should multiply 7 to this map.
    

    Quantitative results are computed from matlab. (e.g., evaluation on the RTF dataset).

    • Options
      • --mode: The name of a model to test. The logging folder named with the [mode] will be created as [LOG_ROOT]/[mode]/. Default: DMENet_BDCS
      • --test_set: The name of a dataset to evaluate. CUHK | RTF0 | RTF1 | RTF1_6 | random. Default: CUHK
        • The folder structure can be modified in the function get_eval_path(..) in ./config.py.
        • random is for testing models with any images, which should be placed as [DATASET_ROOT]/random/*.[jpg|png].
  • Check out the evaluation code for the RTF dataset, and the deconvolution code.

Training & testing the network

  • Train the network by:

    python main.py --is_train --mode [mode]

    Note:

    • If you train DMENet with newly generated SYNDOF dataset from this repo, comment this line and uncomment this line before the training.
  • Test the network by:

    python main.py --mode [mode] --test_set [test_set]
    • arguments
      • --mode: The name of a model to train. The logging folder named with the [mode] will be created as [LOG_ROOT]/[mode]/. Default: DMENet_BDCS
      • --is_pretrain: Pretrain the network with the MSE loss (True | False). Default: False
      • --delete_log: Deletes [LOG_ROOT]/[mode]/* before training begins (True | False). Default: False

Contact

Open an issue for any inquiries. You may also have contact with [email protected]

Related Links

  • CVPR 2021: Iterative Filter Adaptive Network for Single Image Defocus Deblurring [paper][code]
  • ICCV 2021: Single Image Defocus Deblurring Using Kernel-Sharing Parallel Atrous Convolutions [paper][code]
  • SYNDOF dataset generation repo [link]

License

License CC BY-NC
This software is being made available under the terms in the LICENSE file. Any exemptions to these terms require a license from the Pohang University of Science and Technology.

Citation

If you find this code useful, please consider citing:

@InProceedings{Lee2019DMENet,
    author    = {Junyong Lee and Sungkil Lee and Sunghyun Cho and Seungyong Lee},
    title     = {Deep Defocus Map Estimation Using Domain Adaptation},
    booktitle = {Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
    year      = {2019}
}
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].