All Projects → jingjin25 → LFSSR-SAS-PyTorch

jingjin25 / LFSSR-SAS-PyTorch

Licence: other
Repository for "Light Field Spatial Super-resolution Using Deep Efficient Spatial-Angular Separable Convolution" , TIP 2018

Programming Languages

python
139335 projects - #7 most used programming language
matlab
3953 projects
shell
77523 projects

Projects that are alternatives of or similar to LFSSR-SAS-PyTorch

Srgan
Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network
Stars: ✭ 2,641 (+11904.55%)
Mutual labels:  super-resolution
SRDenseNet-pytorch
SRDenseNet-pytorch(ICCV_2017)
Stars: ✭ 113 (+413.64%)
Mutual labels:  super-resolution
AI-Lossless-Zoomer
AI无损放大工具
Stars: ✭ 940 (+4172.73%)
Mutual labels:  super-resolution
SRCNN-PyTorch
Pytorch framework can easily implement srcnn algorithm with excellent performance
Stars: ✭ 48 (+118.18%)
Mutual labels:  super-resolution
ESPCN-PyTorch
A PyTorch implementation of ESPCN based on CVPR 2016 paper Real-Time Single Image and Video Super-Resolution Using an Efficient Sub-Pixel Convolutional Neural Network.
Stars: ✭ 33 (+50%)
Mutual labels:  super-resolution
MLSR
Source code for ECCV2020 "Fast Adaptation to Super-Resolution Networks via Meta-Learning"
Stars: ✭ 59 (+168.18%)
Mutual labels:  super-resolution
Master Thesis Bayesiancnn
Master Thesis on Bayesian Convolutional Neural Network using Variational Inference
Stars: ✭ 222 (+909.09%)
Mutual labels:  super-resolution
tf-perceptual-eusr
A TensorFlow-based image super-resolution model considering both quantitative and perceptual quality
Stars: ✭ 44 (+100%)
Mutual labels:  super-resolution
CSSR
Crack Segmentation for Low-Resolution Images using Joint Learning with Super-Resolution (CSSR) was accepted to international conference on MVA2021 (oral), and selected for the Best Practical Paper Award.
Stars: ✭ 50 (+127.27%)
Mutual labels:  super-resolution
WaifuLite
Super Resolution for Anime image, lightweight implementation
Stars: ✭ 20 (-9.09%)
Mutual labels:  super-resolution
TEGAN
Generative Adversarial Network (GAN) for physically realistic enrichment of turbulent flow fields
Stars: ✭ 60 (+172.73%)
Mutual labels:  super-resolution
Jalali-Lab-Implementation-of-RAISR
Implementation of RAISR (Rapid and Accurate Image Super Resolution) algorithm in Python 3.x by Jalali Laboratory at UCLA. The implementation presented here achieved performance results that are comparable to that presented in Google's research paper (with less than ± 0.1 dB in PSNR). Just-in-time (JIT) compilation employing JIT numba is used to …
Stars: ✭ 118 (+436.36%)
Mutual labels:  super-resolution
pytorch-gans
PyTorch implementation of GANs (Generative Adversarial Networks). DCGAN, Pix2Pix, CycleGAN, SRGAN
Stars: ✭ 21 (-4.55%)
Mutual labels:  super-resolution
Image Super Resolution
🔎 Super-scale your images and run experiments with Residual Dense and Adversarial Networks.
Stars: ✭ 3,293 (+14868.18%)
Mutual labels:  super-resolution
Single-Image-Example-Based-Super-Resolution
Single image example-based super resolution. Improves the spatial and temporal resolution of an image using a direct mapping of LR HR patch pairs. C++, openCV.
Stars: ✭ 33 (+50%)
Mutual labels:  super-resolution
Zoom Learn Zoom
computational zoom from raw sensor data
Stars: ✭ 224 (+918.18%)
Mutual labels:  super-resolution
MSG-Net
Depth Map Super-Resolution by Deep Multi-Scale Guidance, ECCV 2016
Stars: ✭ 76 (+245.45%)
Mutual labels:  super-resolution
colibri-vr-unity-package
This is the Unity package for COLIBRI VR, the Core Open Lab on Image-Based Rendering Innovation for Virtual Reality.
Stars: ✭ 41 (+86.36%)
Mutual labels:  light-fields
FB
Multi-frame super-resolution via sub-pixel.
Stars: ✭ 55 (+150%)
Mutual labels:  super-resolution
tensorflow-bicubic-downsample
tf.image.resize_images has aliasing when downsampling and does not have gradients for bicubic mode. This implementation fixes those problems.
Stars: ✭ 23 (+4.55%)
Mutual labels:  super-resolution

LFSSR-SAS-PyTorch

PyTorch implementation of TIP 2018 paper: "Light Field Spatial Super-resolution Using Deep Efficient Spatial-Angular Separable Convolution". You can find the original MATLAB code from here.

Usage

Dependencies

  • Python 3.6
  • PyTorch 1.0

Dataset

We provide MATLAB code for preparing the training and testing data. Please first download light field datasets, and put them into corresponding folders in LFData.

Note: New training data were generated by bicubic downsampling, while the original training data were generated by spatially blurring with a box filter followed by decimation.

Demo

usage: demo.py [-h] [--model_path MODEL_PATH] [--scale SCALE]
               [--test_dataset TEST_DATASET] [--angular_num ANGULAR_NUM]
               [--layer_num LAYER_NUM] [--save_img SAVE_IMG] 
             
optional arguments:  
  -h, --help          Show this help message and exit  
  --model_path        Model path. Default=pretrained_models/model_2x.pth  
  --scale             SR factor  
  --test_dataset      Dataset for test  
  --angular_num       Size of one angular dim. Default=7.  
  --layer_num         Number of SAS layers. Default=6.  
  --save_img          Save image or not  

An example of usage is shown as follows:

python demo.py --model_path pretrained_models/model_2x.pth --test_dataset HCI --scale 2  --save_img 1

Note: We provide 2 pre-trained models for 2x and 4x SR, respectively. There are some differences from the original MATLAB pre-trained models:

  • new models were trained on a hybrid dataset containing both synthetic and real-world light field images, while the original ones were trained only on real-world images captured by a Lytro Illum camera. The lists of datasets used to train the new models can be found in LFData/dataset_train/trainList_datasetname.txt;
  • new models were trained for light fields with the angular resolution of 7x7, while the original ones were trained for 8x8 light fields;
  • and new models used 6 layers of spatial-angular convolutions, while the original ones used 10 layers.

Training

An example of training your own model is shown as follows:

python train.py --dataset all --scale 2 --layer_num 6 --angular_num 7 --lr 1e-4

Testing

An example of testing one epoch of your trained model is shown as follows:

python test.py --train_dataset all --test_dataset HCI --scale 2 --layer_num 6 --angular_num 7 --lr 1e-4 --epoch 500 --save_img 1
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].