All Projects → gongenhao → Gancs

gongenhao / Gancs

Licence: mit
Compressed Sensing MRI based on Deep Generative Adversarial Network

Projects that are alternatives of or similar to Gancs

Imageclassification
Deep Learning: Image classification, feature visualization and transfer learning with Keras
Stars: ✭ 83 (+0%)
Mutual labels:  jupyter-notebook
Econml
ALICE (Automated Learning and Intelligence for Causation and Economics) is a Microsoft Research project aimed at applying Artificial Intelligence concepts to economic decision making. One of its goals is to build a toolkit that combines state-of-the-art machine learning techniques with econometrics in order to bring automation to complex causal inference problems. To date, the ALICE Python SDK (econml) implements orthogonal machine learning algorithms such as the double machine learning work of Chernozhukov et al. This toolkit is designed to measure the causal effect of some treatment variable(s) t on an outcome variable y, controlling for a set of features x.
Stars: ✭ 1,238 (+1391.57%)
Mutual labels:  jupyter-notebook
Deep Knowledge Tracing Plus
It is the DKT+ model implemented in python3 and tensorflow1.2
Stars: ✭ 83 (+0%)
Mutual labels:  jupyter-notebook
Coronabr
Série histórica dos dados sobre COVID-19, a partir de informações do Ministério da Saúde
Stars: ✭ 83 (+0%)
Mutual labels:  jupyter-notebook
Amazon Sagemaker Script Mode
Amazon SageMaker examples for prebuilt framework mode containers, a.k.a. Script Mode, and more (BYO containers and models etc.)
Stars: ✭ 82 (-1.2%)
Mutual labels:  jupyter-notebook
Platzidata
Stars: ✭ 83 (+0%)
Mutual labels:  jupyter-notebook
Tensorflow Demo
Local AI demo and distributed AI demo using TensorFlow
Stars: ✭ 82 (-1.2%)
Mutual labels:  jupyter-notebook
Jupyter case studies
Inference case studies in jupyter
Stars: ✭ 83 (+0%)
Mutual labels:  jupyter-notebook
Mlcourse
生命情報の機械学習入門(新学術領域「先進ゲノム支援」中級講習会資料)
Stars: ✭ 83 (+0%)
Mutual labels:  jupyter-notebook
Embeddings reproduction
Stars: ✭ 83 (+0%)
Mutual labels:  jupyter-notebook
Machine Learning Portfolio
Machine learning portfolio
Stars: ✭ 83 (+0%)
Mutual labels:  jupyter-notebook
Rsn
Learning to Exploit Long-term Relational Dependencies in Knowledge Graphs, ICML 2019
Stars: ✭ 83 (+0%)
Mutual labels:  jupyter-notebook
Deepembeding
图像检索和向量搜索,similarity learning,compare deep metric and deep-hashing applying in image retrieval
Stars: ✭ 83 (+0%)
Mutual labels:  jupyter-notebook
Animecolordeoldify
Colorise Anime/Manga Sketches with DeOldify
Stars: ✭ 83 (+0%)
Mutual labels:  jupyter-notebook
Logomaker
Software for the visualization of sequence-function relationships
Stars: ✭ 83 (+0%)
Mutual labels:  jupyter-notebook
Ydf Recsys2015 Challenge
Solution of RecSys Challenge 2015
Stars: ✭ 82 (-1.2%)
Mutual labels:  jupyter-notebook
Tensorflow Tutorials
TensorFlow Tutorials with YouTube Videos
Stars: ✭ 8,919 (+10645.78%)
Mutual labels:  jupyter-notebook
Torchtexttutorial
A short tutorial for Torchtext, the NLP-specific add-on for Pytorch.
Stars: ✭ 83 (+0%)
Mutual labels:  jupyter-notebook
Carvana Challenge
My repository for the Carvana Image Masking Challenge
Stars: ✭ 83 (+0%)
Mutual labels:  jupyter-notebook
Spacenet building detection
Project to train/test convolutional neural networks to extract buildings from SpaceNet satellite imageries.
Stars: ✭ 83 (+0%)
Mutual labels:  jupyter-notebook

GANCS

Compressed Sensing MRI based on Generative Adversarial Network

Introduction

First Authors:

Morteza Mardani, Enhao Gong

Arxiv Paper

Deep Generative Adversarial Networks for Compressed Sensing Automates MRI Arxiv Paper: https://arxiv.org/abs/1706.00051

References:

The basic code base is derived from super resolution github repo. https://github.com/david-gpu/srez

https://github.com/david-gpu/srez https://arxiv.org/abs/1609.04802 https://github.com/carpedm20/DCGAN-tensorflow http://wiseodd.github.io/techblog/2017/03/02/least-squares-gan/ http://wiseodd.github.io/techblog/2017/02/04/wasserstein-gan/

Method

Deep Generative Adversarial Networks for CS MRI

Magnetic resonance imaging (MRI) suffers from aliasing artifacts when it is highly undersampled for fast imaging. Conventional CS MRI reconstruction uses regularized iterative reconstruction based on pre-defined sparsity transform, which usually include time-consuming iterative optimization and may result in undesired artifacts such as oversmoothing. Here we propose a novel CS framework that permeates benefits from deep learning and generative adversarial networks (GAN) to modeling a manifold of MR images from historical patients. Extensive evaluations on a large MRI datasets of pediatric pateints show it results in superior perforamnce, retrieves image with improved quality and finer details relative to conventional CS and pixel-wise deep learning schemes.

Undersampling

1D undersampling

1D undersampling is generated using the variable density distribution. R_factor defines the desired reduction factor, which controls how many samples to randomly pick. R_alpha defines the decay of VD distribution with formula p=x^alpha. R_seed defines the random seed used, for negative values there is no fixed undersampling pattern.

1D/2D undersampling

sampling_pattern can be a path to .mat file for specific 1D/2D undersampling mask

Generator Model

Several models are explored including ResNet-ish models from super-resolution paper and encoder-decoder models.

Descriminator Model

Currently we are using 4*(Conv-BN-RELU-POOL)+2*(CONV-BN-RELU)+CONV+MEAN+softmax (for logloss)

descriminator related loss

We have been exploring different loss functions for GAN, including:

  • log-loss
  • LS loss (better than log-loss, use as default, easy to tune and optimize)
  • Cycle-GAN/WGAN loss (todo)

Loss formulation

Loss is a mixed combination with: 1) Data consistency loss, 2) pixel-wise MSE/L1/L2 loss and 3) LS-GAN loss

FLAGS.gene_log_factor = 0 # log loss vs least-square loss

FLAGS.gene_dc_factor = 0.9 # data-consistency (kspace) loss vs generator loss

FLAGS.gene_mse_factor = 0.001 # simple MSE loss originally for forward-passing model vs GAN loss GAN loss = generator loss + discriminator loss

gene_fool_loss = FLAGS.gene_log_factor * gene_log_loss + (1-FLAGS.gene_log_factor) * gene_LS_loss

gene_non_mse_loss = FLAGS.gene_dc_factor * gene_dc_loss + (1-FLAGS.gene_dc_factor) * gene_fool_loss

gene_loss = FLAGS.gene_mse_factor * gene_mse_loss + (1- FLAGS.gene_mse_factor) * gene_non_mse_loss

Dataset

Dataset is parsed to pngs saved in specific folders

  • Phantom dataset
  • Knee dataset
  • DCE dataset

Results

Multiple results are exported while traning

  • loss changes
  • test results (zero-fill VS Recon VS Ref) saved to png for each epoch
  • some of the train results are exported to png, WARNING, there was a memory bug before when we try to export all train results
  • layers (some layers are skipped) of generator and detectors are exported into json after each epoch.

Code structures

Files

  • srez_main.py handles commendline arguments
  • srez_input.py loads from input files, conducting undersampling or loading/extracting the undersampled patterns,
  • srez_model.py defines Generator and Discriminator models
  • srez_train.py trains model in batch and export checkpoints and results for progress summaries

Usage example:

Training example

(currently working on t2) python srez_main.py --dataset_input /home/enhaog/GANCS/srez/dataset_MRI/phantom --batch_size 8 --run train --summary_period 123 --sample_size 256 --train_time 10 --train_dir train_save_all --R_factor 4 --R_alpha 3 --R_seed 0

(currently working on t2 for DCE) python srez_main.py --run train --dataset_input /home/enhaog/GANCS/srez/dataset_MRI/abdominal_DCE --sample_size 200 --sample_size_y 100 --sampling_pattern /home/enhaog/GANCS/srez/dataset_MRI/sampling_pattern_DCE/mask_2dvardesnity_radiaview_4fold.mat --batch_size 4 --summary_period 125 --sample_test 32 --sample_train 10000 --train_time 200 --train_dir train_DCE_test

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].