All Projects → lzx1413 → Pytorchssd

lzx1413 / Pytorchssd

Licence: mit
pytorch version of SSD and it's enhanced methods such as RFBSSD,FSSD and RefineDet

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Pytorchssd

zenfs
ZenFS is a storage backend for RocksDB that enables support for ZNS SSDs and SMR HDDs.
Stars: ✭ 63 (-90.96%)
Mutual labels:  ssd
Ssd.tensorflow
State-of-the-art Single Shot MultiBox Detector in Pure TensorFlow, QQ Group: 758790869
Stars: ✭ 334 (-52.08%)
Mutual labels:  ssd
Ssd.pytorch
A PyTorch Implementation of Single Shot MultiBox Detector
Stars: ✭ 4,499 (+545.48%)
Mutual labels:  ssd
Mmdetection To Tensorrt
convert mmdetection model to tensorrt, support fp16, int8, batch input, dynamic shape etc.
Stars: ✭ 262 (-62.41%)
Mutual labels:  ssd
Fastmot
High-performance multiple object tracking based on YOLO, Deep SORT, and optical flow
Stars: ✭ 284 (-59.25%)
Mutual labels:  ssd
Pytorchnethub
项目注释+论文复现+算法竞赛
Stars: ✭ 341 (-51.08%)
Mutual labels:  ssd
PytorchSSD-HarDNet
achieves 35.1 mAP while as fast as SSD-vgg16
Stars: ✭ 26 (-96.27%)
Mutual labels:  ssd
Ssds.pytorch
Repository for Single Shot MultiBox Detector and its variants, implemented with pytorch, python3.
Stars: ✭ 570 (-18.22%)
Mutual labels:  ssd
Crystaldiskinfo
CrystalDiskInfo
Stars: ✭ 293 (-57.96%)
Mutual labels:  ssd
Freemove
Move directories without breaking shortcuts or installations
Stars: ✭ 443 (-36.44%)
Mutual labels:  ssd
Object Detection Api Tensorflow
Object Detection API Tensorflow
Stars: ✭ 267 (-61.69%)
Mutual labels:  ssd
Ssd Pytorch
SSD目标检测算法(Single Shot MultiBox Detector)(简单,明了,易用,全中文注释,单机多卡训练,视频检测)( If you train the model on a single computer and mutil GPU, this program will be your best choice , easier to use and easier to understand )
Stars: ✭ 276 (-60.4%)
Mutual labels:  ssd
Ssd Tensorflow
Single Shot MultiBox Detector in TensorFlow
Stars: ✭ 4,066 (+483.36%)
Mutual labels:  ssd
powersave
Linux power save settings, compatible with systemd
Stars: ✭ 79 (-88.67%)
Mutual labels:  ssd
Redis source annotation
redis 3.2.8 的源码注释
Stars: ✭ 452 (-35.15%)
Mutual labels:  ssd
ssd detectors
SSD-based object and text detection with Keras, SSD, DSOD, TextBoxes, SegLink, TextBoxes++, CRNN
Stars: ✭ 295 (-57.68%)
Mutual labels:  ssd
Rectlabel Support
RectLabel - An image annotation tool to label images for bounding box object detection and segmentation.
Stars: ✭ 338 (-51.51%)
Mutual labels:  ssd
Sirix
SirixDB is a temporal, evolutionary database system, which uses an accumulate only approach. It keeps the full history of each resource. Every commit stores a space-efficient snapshot through structural sharing. It is log-structured and never overwrites data. SirixDB uses a novel page-level versioning approach called sliding snapshot.
Stars: ✭ 638 (-8.46%)
Mutual labels:  ssd
Ssd tensorflow traffic sign detection
Implementation of Single Shot MultiBox Detector in TensorFlow, to detect and classify traffic signs
Stars: ✭ 459 (-34.15%)
Mutual labels:  ssd
Mobilenetv2 Ssdlite
Caffe implementation of SSD and SSDLite detection on MobileNetv2, converted from tensorflow.
Stars: ✭ 435 (-37.59%)
Mutual labels:  ssd

Pytorch SSD Series

Pytorch 4.1 is suppoted on branch 0.4 now.

Support Arc:

VOC2007 Test

System mAP FPS (Titan X Maxwell)
Faster R-CNN (VGG16) 73.2 7
YOLOv2 (Darknet-19) 78.6 40
R-FCN (ResNet-101) 80.5 9
SSD300* (VGG16) 77.2 46
SSD512* (VGG16) 79.8 19
RFBNet300 (VGG16) 80.5 83
RFBNet512 (VGG16) 82.2 38
SSD300 (VGG) 77.8 150 (1080Ti)
FSSD300 (VGG) 78.8 120 (1080Ti)

COCO

System test-dev mAP Time (Titan X Maxwell)
Faster R-CNN++ (ResNet-101) 34.9 3.36s
YOLOv2 (Darknet-19) 21.6 25ms
SSD300* (VGG16) 25.1 22ms
SSD512* (VGG16) 28.8 53ms
RetinaNet500 (ResNet-101-FPN) 34.4 90ms
RFBNet300 (VGG16) 29.9 15ms*
RFBNet512 (VGG16) 33.8 30ms*
RFBNet512-E (VGG16) 34.4 33ms*
SSD512 (HarDNet68) 31.7 TBD (12.9ms**)
SSD512 (HarDNet85) 35.1 TBD (15.9ms**)
RFBNet512 (HarDNet68) 33.9 TBD (16.7ms**)
RFBNet512 (HarDNet85) 36.8 TBD (19.3ms**)

Note: * The speed here is tested on the newest pytorch and cudnn version (0.2.0 and cudnnV6), which is obviously faster than the speed reported in the paper (using pytorch-0.1.12 and cudnnV5).

Note: ** HarDNet results are measured on Titan V with pytorch 1.0.1 for detection only (NMS is NOT included, which is 13~18ms in general cases). For reference, the measurement of SSD-vgg on the same environment is 15.7ms (also detection only).

MobileNet

System COCO minival mAP #parameters
SSD MobileNet 19.3 6.8M
RFB MobileNet 20.7* 7.4M

*: slightly better than the original ones in the paper (20.5).

Contents

  1. Installation
  2. Datasets
  3. Training
  4. Evaluation
  5. Models

Installation

  • Install PyTorch-0.2.0-0.3.1 by selecting your environment on the website and running the appropriate command.
  • Clone this repository. This repository is mainly based onRFBNet, ssd.pytorch and Chainer-ssd, a huge thank to them.
    • Note: We currently only support Python 3+.
  • Compile the nms and coco tools:
./make.sh

Note*: Check you GPU architecture support in utils/build.py, line 131. Default is:

'nvcc': ['-arch=sm_52',
  • Install pyinn for MobileNet backbone:
pip install git+https://github.com/szagoruyko/[email protected]
  • Then download the dataset by following the instructions below and install opencv.
conda install opencv

Note: For training, we currently support VOC and COCO.

Datasets

To make things easy, we provide simple VOC and COCO dataset loader that inherits torch.utils.data.Dataset making it fully compatible with the torchvision.datasets API.

VOC Dataset

Download VOC2007 trainval & test
# specify a directory for dataset to be downloaded into, else default is ~/data/
sh data/scripts/VOC2007.sh # <directory>
Download VOC2012 trainval
# specify a directory for dataset to be downloaded into, else default is ~/data/
sh data/scripts/VOC2012.sh # <directory>

COCO Dataset

Install the MS COCO dataset at /path/to/coco from official website, default is ~/data/COCO. Following the instructions to prepare minival2014 and valminusminival2014 annotations. All label files (.json) should be under the COCO/annotations/ folder. It should have this basic structure

$COCO/
$COCO/cache/
$COCO/annotations/
$COCO/images/
$COCO/images/test2015/
$COCO/images/train2014/
$COCO/images/val2014/

UPDATE: The current COCO dataset has released new train2017 and val2017 sets which are just new splits of the same image sets.

Training

mkdir weights
cd weights
wget https://s3.amazonaws.com/amdegroot-models/vgg16_reducedfc.pth
  • To train RFBNet using the train script simply specify the parameters listed in train_RFB.py as a flag or manually change them.
python train_test.py -d VOC -v RFB_vgg -s 300 
  • Note:
    • -d: choose datasets, VOC or COCO.
    • -v: choose backbone version, RFB_VGG, RFB_E_VGG or RFB_mobile.
    • -s: image size, 300 or 512.
    • You can pick-up training from a checkpoint by specifying the path as one of the training parameters (again, see train_RFB.py for options)

Evaluation

The test frequency can be found in the train_test.py By default, it will directly output the mAP results on VOC2007 test or COCO minival2014. For VOC2012 test and COCO test-dev results, you can manually change the datasets in the test_RFB.py file, then save the detection results and submitted to the server.

Models

Update (Sep 29, 2019)

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].