Gpt2 ChitchatGPT2 for Chinese chitchat/用于中文闲聊的GPT2模型(实现了DialoGPT的MMI思想)
Stars: ✭ 1,230 (+673.58%)
MinTLMinTL: Minimalist Transfer Learning for Task-Oriented Dialogue Systems
Stars: ✭ 61 (-61.64%)
Distre[ACL 19] Fine-tuning Pre-Trained Transformer Language Models to Distantly Supervised Relation Extraction
Stars: ✭ 75 (-52.83%)
GTSRB Keras STNGerman Traffic Sign Recognition Benchmark, Keras implementation with Spatial Transformer Networks
Stars: ✭ 48 (-69.81%)
DialogptLarge-scale pretraining for dialogue
Stars: ✭ 1,177 (+640.25%)
Neural-Scam-ArtistWeb Scraping, Document Deduplication & GPT-2 Fine-tuning with a newly created scam dataset.
Stars: ✭ 18 (-88.68%)
PresentoPresento - Transformer & Presenter Package for PHP
Stars: ✭ 71 (-55.35%)
amrlibA python library that makes AMR parsing, generation and visualization simple.
Stars: ✭ 107 (-32.7%)
Mixture Of ExpertsA Pytorch implementation of Sparsely-Gated Mixture of Experts, for massively increasing the parameter count of language models
Stars: ✭ 68 (-57.23%)
transformer-lsOfficial PyTorch Implementation of Long-Short Transformer (NeurIPS 2021).
Stars: ✭ 201 (+26.42%)
cometaCorpus of Online Medical EnTities: the cometA corpus
Stars: ✭ 31 (-80.5%)
Deeplearning Nlp ModelsA small, interpretable codebase containing the re-implementation of a few "deep" NLP models in PyTorch. Colab notebooks to run with GPUs. Models: word2vec, CNNs, transformer, gpt.
Stars: ✭ 64 (-59.75%)
gravityUser-space deniable data encryption client.
Stars: ✭ 89 (-44.03%)
Transformer DynetAn Implementation of Transformer (Attention Is All You Need) in DyNet
Stars: ✭ 57 (-64.15%)
HRFormerThis is an official implementation of our NeurIPS 2021 paper "HRFormer: High-Resolution Transformer for Dense Prediction".
Stars: ✭ 357 (+124.53%)
Mojito微信、bilibili大图、长图、gif、视频、自定义view的转场效果,The transition effect of wechat, bilibili large image, long image, GIF, video and custom view
Stars: ✭ 1,068 (+571.7%)
WaveGrad2PyTorch Implementation of Google Brain's WaveGrad 2: Iterative Refinement for Text-to-Speech Synthesis
Stars: ✭ 55 (-65.41%)
Gpt2 FrenchGPT-2 French demo | Démo française de GPT-2
Stars: ✭ 47 (-70.44%)
TokenLabelingPytorch implementation of "All Tokens Matter: Token Labeling for Training Better Vision Transformers"
Stars: ✭ 385 (+142.14%)
SockeyeSequence-to-sequence framework with a focus on Neural Machine Translation based on Apache MXNet
Stars: ✭ 990 (+522.64%)
Jazz transformerTransformer-XL for Jazz music composition. Paper: "The Jazz Transformer on the Front Line: Exploring the Shortcomings of AI-Composed Music through Quantitative Measures", ISMIR 2020
Stars: ✭ 36 (-77.36%)
sisterSImple SenTence EmbeddeR
Stars: ✭ 66 (-58.49%)
Meta EmbMultilingual Meta-Embeddings for Named Entity Recognition (RepL4NLP & EMNLP 2019)
Stars: ✭ 28 (-82.39%)
YOLOSYou Only Look at One Sequence (NeurIPS 2021)
Stars: ✭ 612 (+284.91%)
WitwickyWitwicky: An implementation of Transformer in PyTorch.
Stars: ✭ 21 (-86.79%)
kaggle-champsCode for the CHAMPS Predicting Molecular Properties Kaggle competition
Stars: ✭ 49 (-69.18%)
PaddlenlpNLP Core Library and Model Zoo based on PaddlePaddle 2.0
Stars: ✭ 212 (+33.33%)
Odsc 2020 nlpRepository for ODSC talk related to Deep Learning NLP
Stars: ✭ 20 (-87.42%)
libaiLiBai(李白): A Toolbox for Large-Scale Distributed Parallel Training
Stars: ✭ 284 (+78.62%)
Bert KerasKeras implementation of BERT with pre-trained weights
Stars: ✭ 820 (+415.72%)
Rasa chatbot cnbuilding a chinese dialogue system based on the newest version of rasa(基于最新版本rasa搭建的对话系统)
Stars: ✭ 723 (+354.72%)
Easyflipviewpager📖 The library for creating book and card flip animations in ViewPager in Android
Stars: ✭ 698 (+338.99%)
Transformer-in-TransformerAn Implementation of Transformer in Transformer in TensorFlow for image classification, attention inside local patches
Stars: ✭ 40 (-74.84%)
Laravel ResponderA Laravel Fractal package for building API responses, giving you the power of Fractal with Laravel's elegancy.
Stars: ✭ 673 (+323.27%)
ru-dalleGenerate images from texts. In Russian
Stars: ✭ 1,606 (+910.06%)
SimCalImplementation of ECCV20 paper: the devil is in classification: a simple framework for long-tail object detection and instance segmentation
Stars: ✭ 99 (-37.74%)
ViTs-vs-CNNs[NeurIPS 2021]: Are Transformers More Robust Than CNNs? (Pytorch implementation & checkpoints)
Stars: ✭ 145 (-8.81%)
Awesome Bert NlpA curated list of NLP resources focused on BERT, attention mechanism, Transformer networks, and transfer learning.
Stars: ✭ 567 (+256.6%)
Vision-Language-TransformerVision-Language Transformer and Query Generation for Referring Segmentation (ICCV 2021)
Stars: ✭ 127 (-20.13%)
3DSemanticMapping JINT 2020Repository for the paper "Extending Maps with Semantic and Contextual Object Information for Robot Navigation: a Learning-Based Framework using Visual and Depth Cues"
Stars: ✭ 38 (-76.1%)
Image-CaptionUsing LSTM or Transformer to solve Image Captioning in Pytorch
Stars: ✭ 36 (-77.36%)
text2keywordsTrained T5 and T5-large model for creating keywords from text
Stars: ✭ 53 (-66.67%)
rt-mrcnnReal time instance segmentation with Mask R-CNN, live from webcam feed.
Stars: ✭ 47 (-70.44%)
TRAR-VQA[ICCV 2021] TRAR: Routing the Attention Spans in Transformers for Visual Question Answering -- Official Implementation
Stars: ✭ 49 (-69.18%)
YinThe efficient and elegant JSON:API 1.1 server library for PHP
Stars: ✭ 214 (+34.59%)