All Projects → SwinTransformer → Swin-Transformer-Semantic-Segmentation

SwinTransformer / Swin-Transformer-Semantic-Segmentation

Licence: Apache-2.0 license
This is an official implementation for "Swin Transformer: Hierarchical Vision Transformer using Shifted Windows" on Semantic Segmentation.

Projects that are alternatives of or similar to Swin-Transformer-Semantic-Segmentation

Swin-Transformer
This is an official implementation for "Swin Transformer: Hierarchical Vision Transformer using Shifted Windows".
Stars: ✭ 8,046 (+839.95%)
Mutual labels:  semantic-segmentation, ade20k, swin-transformer
SegFormer
Official PyTorch implementation of SegFormer
Stars: ✭ 1,264 (+47.66%)
Mutual labels:  semantic-segmentation, ade20k
Semantic Segmentation Pytorch
Pytorch implementation for Semantic Segmentation/Scene Parsing on MIT ADE20K dataset
Stars: ✭ 4,151 (+384.93%)
Mutual labels:  semantic-segmentation, ade20k
semantic-segmentation
SOTA Semantic Segmentation Models in PyTorch
Stars: ✭ 464 (-45.79%)
Mutual labels:  semantic-segmentation, ade20k
ObjectNet
PyTorch implementation of "Pyramid Scene Parsing Network".
Stars: ✭ 15 (-98.25%)
Mutual labels:  semantic-segmentation, ade20k
pyconvsegnet
Semantic Segmentation PyTorch code for our paper: Pyramidal Convolution: Rethinking Convolutional Neural Networks for Visual Recognition (https://arxiv.org/pdf/2006.11538.pdf)
Stars: ✭ 32 (-96.26%)
Mutual labels:  semantic-segmentation, ade20k
semantic-segmentation-tensorflow
Semantic segmentation task for ADE20k & cityscapse dataset, based on several models.
Stars: ✭ 84 (-90.19%)
Mutual labels:  semantic-segmentation, ade20k
Indoor-segmentation
Indoor segmentation for robot navigating, which is based on deeplab model in TensorFlow.
Stars: ✭ 105 (-87.73%)
Mutual labels:  semantic-segmentation, ade20k
Mmsegmentation
OpenMMLab Semantic Segmentation Toolbox and Benchmark.
Stars: ✭ 2,875 (+235.86%)
Mutual labels:  semantic-segmentation, swin-transformer
Polyp-Segmentation-using-UNET-in-TensorFlow-2.0
Implementing polyp segmentation using the U-Net and CVC-612 dataset.
Stars: ✭ 52 (-93.93%)
Mutual labels:  semantic-segmentation
SAFNet
[IROS 2021] Implementation of "Similarity-Aware Fusion Network for 3D Semantic Segmentation"
Stars: ✭ 19 (-97.78%)
Mutual labels:  semantic-segmentation
AuxiLearn
Official implementation of Auxiliary Learning by Implicit Differentiation [ICLR 2021]
Stars: ✭ 71 (-91.71%)
Mutual labels:  semantic-segmentation
AdaptationSeg
Curriculum Domain Adaptation for Semantic Segmentation of Urban Scenes, ICCV 2017
Stars: ✭ 128 (-85.05%)
Mutual labels:  semantic-segmentation
map-floodwater-satellite-imagery
This repository focuses on training semantic segmentation models to predict the presence of floodwater for disaster prevention. Models were trained using SageMaker and Colab.
Stars: ✭ 21 (-97.55%)
Mutual labels:  semantic-segmentation
celldetection
Cell Detection with PyTorch.
Stars: ✭ 44 (-94.86%)
Mutual labels:  semantic-segmentation
EDANet
Implementation details for EDANet
Stars: ✭ 34 (-96.03%)
Mutual labels:  semantic-segmentation
LinkNet tensorflow
TensorFlow implementation of LinkNet
Stars: ✭ 16 (-98.13%)
Mutual labels:  semantic-segmentation
food-detection-yolov5
🍔🍟🍗 Food analysis baseline with Theseus. Integrate object detection, image classification and multi-class semantic segmentation. 🍞🍖🍕
Stars: ✭ 68 (-92.06%)
Mutual labels:  semantic-segmentation
MINet
Multi-scale Interaction for Real-time LiDAR Data Segmentation on an Embedded Platform (RA-L)
Stars: ✭ 28 (-96.73%)
Mutual labels:  semantic-segmentation
Video-Swin-Transformer
This is an official implementation for "Video Swin Transformers".
Stars: ✭ 932 (+8.88%)
Mutual labels:  swin-transformer

Swin Transformer for Semantic Segmentaion

This repo contains the supported code and configuration files to reproduce semantic segmentaion results of Swin Transformer. It is based on mmsegmentaion.

Updates

05/11/2021 Models for MoBY are released

04/12/2021 Initial commits

Results and Models

ADE20K

Backbone Method Crop Size Lr Schd mIoU mIoU (ms+flip) #params FLOPs config log model
Swin-T UPerNet 512x512 160K 44.51 45.81 60M 945G config github/baidu github/baidu
Swin-S UperNet 512x512 160K 47.64 49.47 81M 1038G config github/baidu github/baidu
Swin-B UperNet 512x512 160K 48.13 49.72 121M 1188G config github/baidu github/baidu

Notes:

Results of MoBY with Swin Transformer

ADE20K

Backbone Method Crop Size Lr Schd mIoU mIoU (ms+flip) #params FLOPs config log model
Swin-T UPerNet 512x512 160K 44.06 45.58 60M 945G config github/baidu github/baidu

Notes:

  • The learning rate needs to be tuned for best practice.
  • MoBY pre-trained models can be downloaded from MoBY with Swin Transformer.

Usage

Installation

Please refer to get_started.md for installation and dataset preparation.

Inference

# single-gpu testing
python tools/test.py <CONFIG_FILE> <SEG_CHECKPOINT_FILE> --eval mIoU

# multi-gpu testing
tools/dist_test.sh <CONFIG_FILE> <SEG_CHECKPOINT_FILE> <GPU_NUM> --eval mIoU

# multi-gpu, multi-scale testing
tools/dist_test.sh <CONFIG_FILE> <SEG_CHECKPOINT_FILE> <GPU_NUM> --aug-test --eval mIoU

Training

To train with pre-trained models, run:

# single-gpu training
python tools/train.py <CONFIG_FILE> --options model.pretrained=<PRETRAIN_MODEL> [model.backbone.use_checkpoint=True] [other optional arguments]

# multi-gpu training
tools/dist_train.sh <CONFIG_FILE> <GPU_NUM> --options model.pretrained=<PRETRAIN_MODEL> [model.backbone.use_checkpoint=True] [other optional arguments] 

For example, to train an UPerNet model with a Swin-T backbone and 8 gpus, run:

tools/dist_train.sh configs/swin/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k.py 8 --options model.pretrained=<PRETRAIN_MODEL> 

Notes:

  • use_checkpoint is used to save GPU memory. Please refer to this page for more details.
  • The default learning rate and training schedule is for 8 GPUs and 2 imgs/gpu.

Citing Swin Transformer

@article{liu2021Swin,
  title={Swin Transformer: Hierarchical Vision Transformer using Shifted Windows},
  author={Liu, Ze and Lin, Yutong and Cao, Yue and Hu, Han and Wei, Yixuan and Zhang, Zheng and Lin, Stephen and Guo, Baining},
  journal={arXiv preprint arXiv:2103.14030},
  year={2021}
}

Other Links

Image Classification: See Swin Transformer for Image Classification.

Object Detection: See Swin Transformer for Object Detection.

Self-Supervised Learning: See MoBY with Swin Transformer.

Video Recognition, See Video Swin Transformer.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].