All Projects → dronefreak → VisDrone-dataset-python-toolkit

dronefreak / VisDrone-dataset-python-toolkit

Licence: MIT License
This repository provides a basic Pythonic toolkit for the VisDrone-Dataset (2018).

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to VisDrone-dataset-python-toolkit

VisDrone2018
ECCV2018(Challenge-Object Detection in Images)
Stars: ✭ 86 (+186.67%)
Mutual labels:  faster-rcnn, visdrone
object-tracking
Multiple Object Tracking System in Keras + (Detection Network - YOLO)
Stars: ✭ 89 (+196.67%)
Mutual labels:  faster-rcnn
FasterRCNN-pytorch
FasterRCNN is implemented in VGG, ResNet and FPN base.
Stars: ✭ 121 (+303.33%)
Mutual labels:  faster-rcnn
Faster-RCNN-LocNet
A simplified implementation of paper : Improved Localization Accuracy by LocNet for Faster R-CNN Based Text Detection
Stars: ✭ 25 (-16.67%)
Mutual labels:  faster-rcnn
GIouloss CIouloss caffe
Caffe version Generalized & Distance & Complete Iou loss Implementation for Faster RCNN/FPN bbox regression
Stars: ✭ 42 (+40%)
Mutual labels:  faster-rcnn
tf-faster-rcnn
Tensorflow 2 Faster-RCNN implementation from scratch supporting to the batch processing with MobileNetV2 and VGG16 backbones
Stars: ✭ 88 (+193.33%)
Mutual labels:  faster-rcnn
Shadowless
A Fast and Open Source Autonomous Perception System.
Stars: ✭ 29 (-3.33%)
Mutual labels:  faster-rcnn
pytorch-faster-rcnn
No description or website provided.
Stars: ✭ 45 (+50%)
Mutual labels:  faster-rcnn
Depth-VRD
Improving Visual Relation Detection using Depth Maps (ICPR 2020)
Stars: ✭ 33 (+10%)
Mutual labels:  faster-rcnn
Faster RCNN tensorflow
Implementation of Faster RCNN for Vehicle Detection
Stars: ✭ 16 (-46.67%)
Mutual labels:  faster-rcnn
keras-faster-rcnn
keras实现faster rcnn,end2end训练、预测; 持续更新中,见todo... ;欢迎试用、关注并反馈问题
Stars: ✭ 85 (+183.33%)
Mutual labels:  faster-rcnn
Real-Time-Object-Detection-API-using-TensorFlow
A Transfer Learning based Object Detection API that detects all objects in an image, video or live webcam. An SSD model and a Faster R-CNN model was pretrained on Mobile net coco dataset along with a label map in Tensorflow. This model were used to detect objects captured in an image, video or real time webcam. Open CV was used for streaming obj…
Stars: ✭ 50 (+66.67%)
Mutual labels:  faster-rcnn
lightDenseYOLO
A real-time object detection app based on lightDenseYOLO Our lightDenseYOLO is the combination of two components: lightDenseNet as the CNN feature extractor and YOLO v2 as the detection module
Stars: ✭ 20 (-33.33%)
Mutual labels:  faster-rcnn
smd
Simple mmdetection CPU inference
Stars: ✭ 27 (-10%)
Mutual labels:  faster-rcnn
Faster-RCNN-Pytorch-Simple
No description or website provided.
Stars: ✭ 24 (-20%)
Mutual labels:  faster-rcnn
MMTOD
Multi-modal Thermal Object Detector
Stars: ✭ 38 (+26.67%)
Mutual labels:  faster-rcnn
Object-Detection-And-Tracking
Target detection in the first frame and Tracking target by SiamRPN.
Stars: ✭ 33 (+10%)
Mutual labels:  faster-rcnn
frcnn-from-scratch-with-keras
💥Faster R-CNN from scratch written with Keras
Stars: ✭ 157 (+423.33%)
Mutual labels:  faster-rcnn
CFUN
Combining Faster R-CNN and U-net for efficient medical image segmentation
Stars: ✭ 109 (+263.33%)
Mutual labels:  faster-rcnn
faster rcnn
Another pytorch implementation of Faster RCNN.
Stars: ✭ 24 (-20%)
Mutual labels:  faster-rcnn

VisDrone-dataset-python-toolkit

This repository provides a basic Pythonic toolkit for the VisDrone-Dataset (2018). Here, I have converted the existing annotations from the dataset to PASCAL-VOC format (regular .xml files). As of now, only object detection in images has been tackled, I'll be developing Pythonic support for other tasks very soon.

Original Annotations

The original annotations seem to follow this particular style:

 <bbox_left>,<bbox_top>,<bbox_width>,<bbox_height>,<score>,<object_category>,<truncation>,<occlusion>


    Name                                                  Description
-------------------------------------------------------------------------------------------------------------------------------     
 <bbox_left>	     The x coordinate of the top-left corner of the predicted bounding box

 <bbox_top>	     The y coordinate of the top-left corner of the predicted object bounding box

 <bbox_width>	     The width in pixels of the predicted object bounding box

<bbox_height>	     The height in pixels of the predicted object bounding box

   <score>	     The score in the DETECTION file indicates the confidence of the predicted bounding box enclosing 
                     an object instance.
                     The score in GROUNDTRUTH file is set to 1 or 0. 1 indicates the bounding box is considered in evaluation, 
                     while 0 indicates the bounding box will be ignored.
                      
<object_category>    The object category indicates the type of annotated object, (i.e., ignored regions(0), pedestrian(1), 
                     people(2), bicycle(3), car(4), van(5), truck(6), tricycle(7), awning-tricycle(8), bus(9), motor(10), 
                     others(11))
                      
<truncation>	     The score in the DETECTION result file should be set to the constant -1.
                     The score in the GROUNDTRUTH file indicates the degree of object parts appears outside a frame 
                     (i.e., no truncation = 0 (truncation ratio 0%), and partial truncation = 1 (truncation ratio 1% ~ 50%)).
                      
<occlusion>	     The score in the DETECTION file should be set to the constant -1.
                     The score in the GROUNDTRUTH file indicates the fraction of objects being occluded (i.e., no occlusion = 0 
                     (occlusion ratio 0%), partial occlusion = 1 (occlusion ratio 1% ~ 50%), and heavy occlusion = 2 
                     (occlusion ratio 50% ~ 100%)).

Conversion to PASCAL-VOC

So this annotation format is converted to PASCAL-VOC using convertVis_to_xml.py. The directory structure that needs to be followed is mentioned below:

VisDrone2019-DET-train
---annotations
---images

convertVis_to_xml.py

So if you just execute this code (python3 convertVis_to_xml.py) from outside the VisDrone2019-DET-train, it will automatically find all the files and corresponding annotations and create two new folders inside the VisDrone2019-DET-train directory like below:

VisDrone2019-DET-train
---annotations
---images
---images_new
---annotations_new

convertVis_to_xml.py

The folders with _new extension have the files required for further manipulations. A sample visualization is shown below:

Original Image With PASCAL-style bounding boxes

Tensorflow's Object Detection API

The annotations seem to be decent enough. A further confirmation is provided by the following results after training a Faster-RCNN-InceptionV3 object detection algorithm on the customized dataset.

Original Image Detection Results

For replicating these results, please clone this repository and just execute: python3 vis-drone-faster-rcnn.py. This code is meant for static images and will save the output to the same directory. A sample trained model is provided in the inference_graph directory.

For now, the code convertVis_to_xml.py creates two new folders and puts the data in them, but this can be changed by altering the write folders in the code. This should be customized as per needs of the user. Thank you and happy coding !!!

References

https://github.com/VisDrone/VisDrone-Dataset

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].