EnsmallenA header-only C++ library for numerical optimization --
Stars: ✭ 436 (+531.88%)
PyswarmsA research toolkit for particle swarm optimization in Python
Stars: ✭ 742 (+975.36%)
Speech TransformerA PyTorch implementation of Speech Transformer, an End-to-End ASR with Transformer network on Mandarin Chinese.
Stars: ✭ 565 (+718.84%)
Nlp tensorflow projectUse tensorflow to achieve some NLP project, eg: classification chatbot ner attention QAetc.
Stars: ✭ 27 (-60.87%)
Rnn NluA TensorFlow implementation of Recurrent Neural Networks for Sequence Classification and Sequence Labeling
Stars: ✭ 463 (+571.01%)
AttentionsPyTorch implementation of some attentions for Deep Learning Researchers.
Stars: ✭ 39 (-43.48%)
OptimOptimLib: a lightweight C++ library of numerical optimization methods for nonlinear functions
Stars: ✭ 411 (+495.65%)
Text ClassificationImplementation of papers for text classification task on DBpedia
Stars: ✭ 682 (+888.41%)
SimplecvreproductionReproduce simple cv project including attention module, classification, object detection, segmentation, keypoint detection, tracking 😄 etc.
Stars: ✭ 602 (+772.46%)
Fmin Unconstrained function minimization in Javascript
Stars: ✭ 317 (+359.42%)
Isab PytorchAn implementation of (Induced) Set Attention Block, from the Set Transformers paper
Stars: ✭ 21 (-69.57%)
Solid🎯 A comprehensive gradient-free optimization framework written in Python
Stars: ✭ 546 (+691.3%)
Awesome RoboticsA curated list of awesome links and software libraries that are useful for robots.
Stars: ✭ 478 (+592.75%)
Pytorch GatMy implementation of the original GAT paper (Veličković et al.). I've additionally included the playground.py file for visualizing the Cora dataset, GAT embeddings, an attention mechanism, and entropy histograms. I've supported both Cora (transductive) and PPI (inductive) examples!
Stars: ✭ 908 (+1215.94%)
Ban VqaBilinear attention networks for visual question answering
Stars: ✭ 449 (+550.72%)
FluenceA deep learning library based on Pytorch focussed on low resource language research and robustness
Stars: ✭ 54 (-21.74%)
Deep learning nlpKeras, PyTorch, and NumPy Implementations of Deep Learning Architectures for NLP
Stars: ✭ 407 (+489.86%)
AttentionclusterTensorFlow Implementation of "Attention Clusters: Purely Attention Based Local Feature Integration for Video Classification"
Stars: ✭ 33 (-52.17%)
Ojalgooj! Algorithms
Stars: ✭ 336 (+386.96%)
Cppnumericalsolversa lightweight C++17 library of numerical optimization methods for nonlinear functions (Including L-BFGS-B for TensorFlow)
Stars: ✭ 638 (+824.64%)
Dist KerasDistributed Deep Learning, with a focus on distributed training, using Keras and Apache Spark.
Stars: ✭ 613 (+788.41%)
Time AttentionImplementation of RNN for Time Series prediction from the paper https://arxiv.org/abs/1704.02971
Stars: ✭ 52 (-24.64%)
Performer PytorchAn implementation of Performer, a linear attention-based transformer, in Pytorch
Stars: ✭ 546 (+691.3%)
MindseyeNeural Networks in Java 8 with CuDNN and Aparapi
Stars: ✭ 8 (-88.41%)
Pagmo2A C++ platform to perform parallel computations of optimisation tasks (global and local) via the asynchronous generalized island model.
Stars: ✭ 540 (+682.61%)
Yolov4 PytorchThis is a pytorch repository of YOLOv4, attentive YOLOv4 and mobilenet YOLOv4 with PASCAL VOC and COCO
Stars: ✭ 1,070 (+1450.72%)
Punctuator2A bidirectional recurrent neural network model with attention mechanism for restoring missing punctuation in unsegmented text
Stars: ✭ 483 (+600%)
Cell DetrOfficial and maintained implementation of the paper Attention-Based Transformers for Instance Segmentation of Cells in Microstructures [BIBM 2020].
Stars: ✭ 26 (-62.32%)
Chinesenre中文实体关系抽取,pytorch,bilstm+attention
Stars: ✭ 463 (+571.01%)
Biblosa PytorchRe-implementation of Bi-Directional Block Self-Attention for Fast and Memory-Efficient Sequence Modeling (T. Shen et al., ICLR 2018) on Pytorch.
Stars: ✭ 43 (-37.68%)
Mac NetworkImplementation for the paper "Compositional Attention Networks for Machine Reasoning" (Hudson and Manning, ICLR 2018)
Stars: ✭ 444 (+543.48%)
Global Self Attention NetworkA Pytorch implementation of Global Self-Attention Network, a fully-attention backbone for vision tasks
Stars: ✭ 64 (-7.25%)
GansformerGenerative Adversarial Transformers
Stars: ✭ 421 (+510.14%)
Tf Rnn AttentionTensorflow implementation of attention mechanism for text classification tasks.
Stars: ✭ 735 (+965.22%)
Pytorch Original TransformerMy implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pretrained models.
Stars: ✭ 411 (+495.65%)
Neural spEnd-to-end ASR/LM implementation with PyTorch
Stars: ✭ 408 (+491.3%)
Nlp TutorialsSimple implementations of NLP models. Tutorials are written in Chinese on my website https://mofanpy.com
Stars: ✭ 394 (+471.01%)
Ner BertBERT-NER (nert-bert) with google bert https://github.com/google-research.
Stars: ✭ 339 (+391.3%)
GaftA Genetic Algorithm Framework in Python
Stars: ✭ 651 (+843.48%)
Transformer TensorflowTensorFlow implementation of 'Attention Is All You Need (2017. 6)'
Stars: ✭ 319 (+362.32%)
Attentive Neural Processesimplementing "recurrent attentive neural processes" to forecast power usage (w. LSTM baseline, MCDropout)
Stars: ✭ 33 (-52.17%)
Deeplearning Nlp ModelsA small, interpretable codebase containing the re-implementation of a few "deep" NLP models in PyTorch. Colab notebooks to run with GPUs. Models: word2vec, CNNs, transformer, gpt.
Stars: ✭ 64 (-7.25%)
DefactonlpDeFactoNLP: An Automated Fact-checking System that uses Named Entity Recognition, TF-IDF vector comparison and Decomposable Attention models.
Stars: ✭ 30 (-56.52%)
VadVoice activity detection (VAD) toolkit including DNN, bDNN, LSTM and ACAM based VAD. We also provide our directly recorded dataset.
Stars: ✭ 622 (+801.45%)