All Projects → bupt-ai-cz → CAC-UNet-DigestPath2019

bupt-ai-cz / CAC-UNet-DigestPath2019

Licence: other
1st to MICCAI DigestPath2019 challenge (https://digestpath2019.grand-challenge.org/Home/) on colonoscopy tissue segmentation and classification task. (MICCAI 2019) https://teacher.bupt.edu.cn/zhuchuang/en/index.htm

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to CAC-UNet-DigestPath2019

Deep-Unsupervised-Domain-Adaptation
Pytorch implementation of four neural network based domain adaptation techniques: DeepCORAL, DDC, CDAN and CDAN+E. Evaluated on benchmark dataset Office31.
Stars: ✭ 50 (-39.76%)
Mutual labels:  domain-adaptation
IAST-ECCV2020
IAST: Instance Adaptive Self-training for Unsupervised Domain Adaptation (ECCV 2020) https://teacher.bupt.edu.cn/zhuchuang/en/index.htm
Stars: ✭ 84 (+1.2%)
Mutual labels:  domain-adaptation
histopathologic cancer detector
CNN histopathologic tumor identifier.
Stars: ✭ 26 (-68.67%)
Mutual labels:  cancer-detection
Domain-Consensus-Clustering
[CVPR2021] Domain Consensus Clustering for Universal Domain Adaptation
Stars: ✭ 85 (+2.41%)
Mutual labels:  domain-adaptation
multichannel-semseg-with-uda
Multichannel Semantic Segmentation with Unsupervised Domain Adaptation
Stars: ✭ 19 (-77.11%)
Mutual labels:  domain-adaptation
DAS
Code and datasets for EMNLP2018 paper ‘‘Adaptive Semi-supervised Learning for Cross-domain Sentiment Classification’’.
Stars: ✭ 48 (-42.17%)
Mutual labels:  domain-adaptation
DAOSL
Implementation of Domain Adaption in One-Shot Learning
Stars: ✭ 14 (-83.13%)
Mutual labels:  domain-adaptation
CrossNER
CrossNER: Evaluating Cross-Domain Named Entity Recognition (AAAI-2021)
Stars: ✭ 87 (+4.82%)
Mutual labels:  domain-adaptation
kits19-challenge
Kidney Tumor Segmentation Challenge 2019
Stars: ✭ 44 (-46.99%)
Mutual labels:  miccai-2019
AdaptationSeg
Curriculum Domain Adaptation for Semantic Segmentation of Urban Scenes, ICCV 2017
Stars: ✭ 128 (+54.22%)
Mutual labels:  domain-adaptation
visda2019-multisource
Source code of our submission (Rank 1) for Multi-Source Domain Adaptation task in VisDA-2019
Stars: ✭ 49 (-40.96%)
Mutual labels:  domain-adaptation
Transfer-learning-materials
resource collection for transfer learning!
Stars: ✭ 213 (+156.63%)
Mutual labels:  domain-adaptation
BA3US
code for our ECCV 2020 paper "A Balanced and Uncertainty-aware Approach for Partial Domain Adaptation"
Stars: ✭ 31 (-62.65%)
Mutual labels:  domain-adaptation
Foveation-Segmentation
PyTorch implementation of Foveation for Segmentation of Ultra-High Resolution Images
Stars: ✭ 38 (-54.22%)
Mutual labels:  histopathology-images
DCAN
[AAAI 2020] Code release for "Domain Conditioned Adaptation Network" https://arxiv.org/abs/2005.06717
Stars: ✭ 27 (-67.47%)
Mutual labels:  domain-adaptation
StyleGAN-nada
stylegan-nada.github.io/
Stars: ✭ 1,018 (+1126.51%)
Mutual labels:  domain-adaptation
SFA
Official Implementation of "Exploring Sequence Feature Alignment for Domain Adaptive Detection Transformers"
Stars: ✭ 79 (-4.82%)
Mutual labels:  domain-adaptation
meta-learning-progress
Repository to track the progress in Meta-Learning (MtL), including the datasets and the current state-of-the-art for the most common MtL problems.
Stars: ✭ 26 (-68.67%)
Mutual labels:  domain-adaptation
Transferable-E2E-ABSA
Transferable End-to-End Aspect-based Sentiment Analysis with Selective Adversarial Learning (EMNLP'19)
Stars: ✭ 62 (-25.3%)
Mutual labels:  domain-adaptation
Transformers-Domain-Adaptation
Adapt Transformer-based language models to new text domains
Stars: ✭ 67 (-19.28%)
Mutual labels:  domain-adaptation

Multi-level-colonoscopy-malignant-tissue-detection-with-adversarial-CAC-UNet Tweet

Implementation detail for our paper "Multi-level colonoscopy malignant tissue detection with adversarial CAC-UNet"

DigestPath 2019

The proposed scheme in this paper achieves the best results in MICCAI DigestPath2019 challenge (https://digestpath2019.grand-challenge.org/Home/) on colonoscopy tissue segmentation and classification task.

Dataset

Description of dataset can be found here: https://digestpath2019.grand-challenge.org/Dataset/

To download the the DigestPath2019 dataset, please sign the DATABASE USE AGREEMENT first at here: https://digestpath2019.grand-challenge.org/Download/

If you have problems about downing the dataset, please contact Prof. Hongsheng Li:[email protected]

Image sample:

Envs

  • Pytorch 1.0
  • Python 3+
  • cuda 9.0+

install

$ pip install -r  requirements.txt

apex : Tools for easy mixed precision and distributed training in Pytorch

$ git clone https://github.com/NVIDIA/apex
$ cd apex
$ pip install -v --no-cache-dir --global-option="--cpp_ext" --global-option="--cuda_ext" ./

Dataset

├── data/
│   ├── tissue-train-neg/     
│   ├── tissue-train-pos-v1/

Preprocessing

$ cd code/
$ python preprocessing.py

Training

$ cd code/
$ python train.py --config_file='config/cac-unet-r50.yaml'

Citation

Please cite this paper in your publications if it helps your research:

@article{zhu2021multi,
  title={Multi-level colonoscopy malignant tissue detection with adversarial CAC-UNet},
  author={Zhu, Chuang and Mei, Ke and Peng, Ting and Luo, Yihao and Liu, Jun and Wang, Ying and Jin, Mulan},
  journal={Neurocomputing},
  volume={438},
  pages={165--183},
  year={2021},
  publisher={Elsevier}
}

About the multi-level adversarial segmentation part, you can read our ICASSP paper for more details:

@inproceedings{mei2020cross,
  title={Cross-stained segmentation from renal biopsy images using multi-level adversarial learning},
  author={Mei, Ke and Zhu, Chuang and Jiang, Lei and Liu, Jun and Qiao, Yuanyuan},
  booktitle={ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)},
  pages={1424--1428},
  year={2020},
  organization={IEEE}
}

Author

Ke Mei, Ting Peng, Chuang Zhu

If you have any questions, you can contact me directly.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].