All Projects → Wovchena → text-detection-fots.pytorch

Wovchena / text-detection-fots.pytorch

Licence: other
FOTS text detection branch reimplementation, hmean: 83.3%

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to text-detection-fots.pytorch

East icpr
Forked from argman/EAST for the ICPR MTWI 2018 CHALLENGE
Stars: ✭ 154 (+92.5%)
Mutual labels:  text-detection
East
A tensorflow implementation of EAST text detector
Stars: ✭ 2,804 (+3405%)
Mutual labels:  text-detection
FaceAttr
CVPR2018 Face Super-resolution with supplementary Attributes
Stars: ✭ 18 (-77.5%)
Mutual labels:  cvpr2018
Adelaidet
AdelaiDet is an open source toolbox for multiple instance-level detection and recognition tasks.
Stars: ✭ 2,565 (+3106.25%)
Mutual labels:  text-detection
React Native Text Detector
Text Detector from image for react native using firebase MLKit on android and Tesseract on iOS
Stars: ✭ 194 (+142.5%)
Mutual labels:  text-detection
AE TextSpotter
No description or website provided.
Stars: ✭ 68 (-15%)
Mutual labels:  text-detection
Tedeval
TedEval: A Fair Evaluation Metric for Scene Text Detectors
Stars: ✭ 143 (+78.75%)
Mutual labels:  text-detection
IDN-pytorch
paper implement : Fast and Accurate Single Image Super-Resolution via Information Distillation Network
Stars: ✭ 40 (-50%)
Mutual labels:  cvpr2018
Ocr.pytorch
A pure pytorch implemented ocr project including text detection and recognition
Stars: ✭ 196 (+145%)
Mutual labels:  text-detection
doctr
docTR (Document Text Recognition) - a seamless, high-performing & accessible library for OCR-related tasks powered by Deep Learning.
Stars: ✭ 1,409 (+1661.25%)
Mutual labels:  text-detection
Text Detection
Text detection with mainly MSER and SWT
Stars: ✭ 167 (+108.75%)
Mutual labels:  text-detection
Awesome Deep Text Detection Recognition
A curated list of resources for text detection/recognition (optical character recognition ) with deep learning methods.
Stars: ✭ 2,282 (+2752.5%)
Mutual labels:  text-detection
glimpse clouds
Pytorch implementation of the paper "Glimpse Clouds: Human Activity Recognition from Unstructured Feature Points", F. Baradel, C. Wolf, J. Mille , G.W. Taylor, CVPR 2018
Stars: ✭ 30 (-62.5%)
Mutual labels:  cvpr2018
Craft Pytorch
Official implementation of Character Region Awareness for Text Detection (CRAFT)
Stars: ✭ 2,220 (+2675%)
Mutual labels:  text-detection
Real-time-Text-Detection-DBNet
PyTorch re-implementation of ''Real-time Scene Text Detection with Differentiable Binarization'' (AAAI 2020)
Stars: ✭ 85 (+6.25%)
Mutual labels:  text-detection
Derpn
A novel region proposal network for more general object detection ( including scene text detection ).
Stars: ✭ 155 (+93.75%)
Mutual labels:  text-detection
google-vision-sampler
Code examples for Google Vision API.
Stars: ✭ 47 (-41.25%)
Mutual labels:  text-detection
vietnamese-ocr-toolbox
A toolbox for Vietnamese Optical Character Recognition.
Stars: ✭ 26 (-67.5%)
Mutual labels:  text-detection
shinTB
Textboxes : Image Text Detection Model : python package (tensorflow)
Stars: ✭ 90 (+12.5%)
Mutual labels:  text-detection
pytorch.ctpn
pytorch, ctpn ,text detection ,ocr,文本检测
Stars: ✭ 123 (+53.75%)
Mutual labels:  text-detection

FOTS: Fast Oriented Text Spotting with a Unified Network text detection branch reimplementation (PyTorch)

Train

  1. Train with SynthText for 9 epochs

    time python3 train.py --train-folder SynthText/ --batch-size 21 --batches-before-train 2

    At this point the result was Epoch 8: 100%|█████████████| 390/390 [08:28<00:00, 1.00it/s, Mean loss=0.98050].

  2. Train with ICDAR15

    Replace a data set in data_set = datasets.SynthText(args.train_folder, datasets.transform) with datasets.ICDAR2015 in train.py and run

    time python3 train.py --train-folder icdar15/ --continue-training --batch-size 21 --batches-before-train 2

    It is expected that the provided --train-folder contains unzipped ch4_training_images and ch4_training_localization_transcription_gt. To avoid saving model at each epoch, the line if True: in train.py can be replaced with if epoch > 60 and epoch % 6 == 0:

    The result was Epoch 582: 100%|█████████████| 48/48 [01:05<00:00, 1.04s/it, Mean loss=0.11290].

Learning rate schedule:

Epoch 175: reducing learning rate of group 0 to 5.0000e-04.

Epoch 264: reducing learning rate of group 0 to 2.5000e-04.

Epoch 347: reducing learning rate of group 0 to 1.2500e-04.

Epoch 412: reducing learning rate of group 0 to 6.2500e-05.

Epoch 469: reducing learning rate of group 0 to 3.1250e-05.

Epoch 525: reducing learning rate of group 0 to 1.5625e-05.

Epoch 581: reducing learning rate of group 0 to 7.8125e-06.

Test

python3 test.py --images-folder ch4_test_images/ --output-folder res/ --checkpoint epoch_582_checkpoint.pt && zip -jmq runs/u.zip res/* && python2 script.py -g=gt.zip -s=runs/u.zip

ch4_training_images and ch4_training_localization_transcription_gt are available in Task 4.4: End to End (2015 edition). script.py and ch4_test_images can be found in My Methods (Script: IoU and test set samples).

It gives Calculated!{"precision": 0.8694968553459119, "recall": 0.7987481945113144, "hmean": 0.8326223337515684, "AP": 0}.

The pretrained models are here: https://drive.google.com/open?id=1xaVshLRrMEkb9LA46IJAZhlapQr3vyY2

test.py has a commented code to visualize results.

Difference with the paper

  1. The model is different compared to what the paper describes. An explanation is in model.py.
  2. The authors of FOTS could not train on clipped words because they also have a recognition branch. The whole word is required to be present on an image to be able to be recognized correctly. This reimplementation has only detection branch and that allows to train on crops of the words.
  3. The paper suggest using some other data sets in addition. Training on SynthText is simplified in this reimplementation.
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].