All Projects → are-16-heads-really-better-than-1 → Similar Projects or Alternatives

540 Open source projects that are alternatives of or similar to are-16-heads-really-better-than-1

Quality-Estimation2
机器翻译子任务-翻译质量评价-在BERT模型后面加上Bi-LSTM进行fine-tuning
Stars: ✭ 31 (-75.78%)
Mutual labels:  bert
Transformer Survey Study
"A survey of Transformer" paper study 👩🏻‍💻🧑🏻‍💻 KoreaUniv. DSBA Lab
Stars: ✭ 166 (+29.69%)
Mutual labels:  transformer
ALBERT-Pytorch
Pytorch Implementation of ALBERT(A Lite BERT for Self-supervised Learning of Language Representations)
Stars: ✭ 214 (+67.19%)
Mutual labels:  bert
LMMS
Language Modelling Makes Sense - WSD (and more) with Contextual Embeddings
Stars: ✭ 79 (-38.28%)
Mutual labels:  bert
zero-administration-inference-with-aws-lambda-for-hugging-face
Zero administration inference with AWS Lambda for 🤗
Stars: ✭ 19 (-85.16%)
Mutual labels:  transformer
transformer-slt
Sign Language Translation with Transformers (COLING'2020, ECCV'20 SLRTP Workshop)
Stars: ✭ 92 (-28.12%)
Mutual labels:  transformer
M3DETR
Code base for M3DeTR: Multi-representation, Multi-scale, Mutual-relation 3D Object Detection with Transformers
Stars: ✭ 47 (-63.28%)
Mutual labels:  transformer
GoEmotions-pytorch
Pytorch Implementation of GoEmotions 😍😢😱
Stars: ✭ 95 (-25.78%)
Mutual labels:  bert
Text and Audio classification with Bert
Text Classification in Turkish Texts with Bert
Stars: ✭ 34 (-73.44%)
Mutual labels:  bert
visualization
a collection of visualization function
Stars: ✭ 189 (+47.66%)
Mutual labels:  transformer
bert quora question pairs
BERT Model Fine-tuning on Quora Questions Pairs
Stars: ✭ 28 (-78.12%)
Mutual labels:  bert
Self-Supervised-Embedding-Fusion-Transformer
The code for our IEEE ACCESS (2020) paper Multimodal Emotion Recognition with Transformer-Based Self Supervised Feature Fusion.
Stars: ✭ 57 (-55.47%)
Mutual labels:  bert
KLUE
📖 Korean NLU Benchmark
Stars: ✭ 420 (+228.13%)
Mutual labels:  bert
Relation-Extraction-Transformer
NLP: Relation extraction with position-aware self-attention transformer
Stars: ✭ 63 (-50.78%)
Mutual labels:  transformer
TadTR
End-to-end Temporal Action Detection with Transformer. [Under review for a journal publication]
Stars: ✭ 55 (-57.03%)
Mutual labels:  transformer
DeepNER
An Easy-to-use, Modular and Prolongable package of deep-learning based Named Entity Recognition Models.
Stars: ✭ 9 (-92.97%)
Mutual labels:  bert
well-classified-examples-are-underestimated
Code for the AAAI 2022 publication "Well-classified Examples are Underestimated in Classification with Deep Neural Networks"
Stars: ✭ 21 (-83.59%)
Mutual labels:  transformer
muse-as-service
REST API for sentence tokenization and embedding using Multilingual Universal Sentence Encoder.
Stars: ✭ 45 (-64.84%)
Mutual labels:  bert
Embedding
Embedding模型代码和学习笔记总结
Stars: ✭ 25 (-80.47%)
Mutual labels:  transformer
FNet-pytorch
Unofficial implementation of Google's FNet: Mixing Tokens with Fourier Transforms
Stars: ✭ 204 (+59.38%)
Mutual labels:  transformer
classy
classy is a simple-to-use library for building high-performance Machine Learning models in NLP.
Stars: ✭ 61 (-52.34%)
Mutual labels:  bert
viewpoint-mining
参考NER,基于BERT的电商评论观点挖掘和情感分析
Stars: ✭ 31 (-75.78%)
Mutual labels:  bert
bert tokenization for java
This is a java version of Chinese tokenization descried in BERT.
Stars: ✭ 39 (-69.53%)
Mutual labels:  bert
contextualSpellCheck
✔️Contextual word checker for better suggestions
Stars: ✭ 274 (+114.06%)
Mutual labels:  bert
max-deeplab
Unofficial implementation of MaX-DeepLab for Instance Segmentation
Stars: ✭ 84 (-34.37%)
Mutual labels:  transformer
KAREN
KAREN: Unifying Hatespeech Detection and Benchmarking
Stars: ✭ 18 (-85.94%)
Mutual labels:  bert
Walk-Transformer
From Random Walks to Transformer for Learning Node Embeddings (ECML-PKDD 2020) (In Pytorch and Tensorflow)
Stars: ✭ 26 (-79.69%)
Mutual labels:  transformer
towhee
Towhee is a framework that is dedicated to making neural data processing pipelines simple and fast.
Stars: ✭ 821 (+541.41%)
Mutual labels:  transformer
mcQA
🔮 Answering multiple choice questions with Language Models.
Stars: ✭ 23 (-82.03%)
Mutual labels:  bert
tf2-transformer-chatbot
Transformer Chatbot in TensorFlow 2 with TPU support.
Stars: ✭ 94 (-26.56%)
Mutual labels:  transformer
bern
A neural named entity recognition and multi-type normalization tool for biomedical text mining
Stars: ✭ 151 (+17.97%)
Mutual labels:  bert
image-classification
A collection of SOTA Image Classification Models in PyTorch
Stars: ✭ 70 (-45.31%)
Mutual labels:  transformer
TorchBlocks
A PyTorch-based toolkit for natural language processing
Stars: ✭ 85 (-33.59%)
Mutual labels:  bert
WSDM-Cup-2019
[ACM-WSDM] 3rd place solution at WSDM Cup 2019, Fake News Classification on Kaggle.
Stars: ✭ 62 (-51.56%)
Mutual labels:  bert
parsbert-ner
🤗 ParsBERT Persian NER Tasks
Stars: ✭ 15 (-88.28%)
Mutual labels:  bert
ai web RISKOUT BTS
국방 리스크 관리 플랫폼 (🏅 국방부장관상/Minister of National Defense Award)
Stars: ✭ 18 (-85.94%)
Mutual labels:  bert
transform-graphql
⚙️ Transformer function to transform GraphQL Directives. Create model CRUD directive for example
Stars: ✭ 23 (-82.03%)
Mutual labels:  transformer
text2text
Text2Text: Cross-lingual natural language processing and generation toolkit
Stars: ✭ 188 (+46.88%)
Mutual labels:  bert
deep-molecular-optimization
Molecular optimization by capturing chemist’s intuition using the Seq2Seq with attention and the Transformer
Stars: ✭ 60 (-53.12%)
Mutual labels:  transformer
SQUAD2.Q-Augmented-Dataset
Augmented version of SQUAD 2.0 for Questions
Stars: ✭ 31 (-75.78%)
Mutual labels:  bert
bert-movie-reviews-sentiment-classifier
Build a Movie Reviews Sentiment Classifier with Google's BERT Language Model
Stars: ✭ 12 (-90.62%)
Mutual labels:  bert
YOLOv5-Lite
🍅🍅🍅YOLOv5-Lite: lighter, faster and easier to deploy. Evolved from yolov5 and the size of model is only 930+kb (int8) and 1.7M (fp16). It can reach 10+ FPS on the Raspberry Pi 4B when the input size is 320×320~
Stars: ✭ 1,230 (+860.94%)
Mutual labels:  transformer
FragmentVC
Any-to-any voice conversion by end-to-end extracting and fusing fine-grained voice fragments with attention
Stars: ✭ 134 (+4.69%)
Mutual labels:  transformer
textwiser
[AAAI 2021] TextWiser: Text Featurization Library
Stars: ✭ 26 (-79.69%)
Mutual labels:  bert
tfbert
基于tensorflow1.x的预训练模型调用,支持单机多卡、梯度累积,XLA加速,混合精度。可灵活训练、验证、预测。
Stars: ✭ 54 (-57.81%)
Mutual labels:  bert
textgo
Text preprocessing, representation, similarity calculation, text search and classification. Let's go and play with text!
Stars: ✭ 33 (-74.22%)
Mutual labels:  bert
BERT-Chinese-Couplet
BERT for Chinese Couplet | BERT用于自动对对联
Stars: ✭ 19 (-85.16%)
Mutual labels:  bert
JD2Skills-BERT-XMLC
Code and Dataset for the Bhola et al. (2020) Retrieving Skills from Job Descriptions: A Language Model Based Extreme Multi-label Classification Framework
Stars: ✭ 33 (-74.22%)
Mutual labels:  bert
TransPose
PyTorch Implementation for "TransPose: Keypoint localization via Transformer", ICCV 2021.
Stars: ✭ 250 (+95.31%)
Mutual labels:  transformer
transformer
A PyTorch Implementation of "Attention Is All You Need"
Stars: ✭ 28 (-78.12%)
Mutual labels:  transformer
kosr
Korean speech recognition based on transformer (트랜스포머 기반 한국어 음성 인식)
Stars: ✭ 25 (-80.47%)
Mutual labels:  transformer
NER-FunTool
本NER项目包含多个中文数据集,模型采用BiLSTM+CRF、BERT+Softmax、BERT+Cascade、BERT+WOL等,最后用TFServing进行模型部署,线上推理和线下推理。
Stars: ✭ 56 (-56.25%)
Mutual labels:  bert
Transformer-in-Transformer
An Implementation of Transformer in Transformer in TensorFlow for image classification, attention inside local patches
Stars: ✭ 40 (-68.75%)
Mutual labels:  transformer
enformer-pytorch
Implementation of Enformer, Deepmind's attention network for predicting gene expression, in Pytorch
Stars: ✭ 146 (+14.06%)
Mutual labels:  transformer
BossNAS
(ICCV 2021) BossNAS: Exploring Hybrid CNN-transformers with Block-wisely Self-supervised Neural Architecture Search
Stars: ✭ 125 (-2.34%)
Mutual labels:  transformer
OpenPrompt
An Open-Source Framework for Prompt-Learning.
Stars: ✭ 1,769 (+1282.03%)
Mutual labels:  transformer
TS-CAM
Codes for TS-CAM: Token Semantic Coupled Attention Map for Weakly Supervised Object Localization.
Stars: ✭ 96 (-25%)
Mutual labels:  transformer
graph-transformer-pytorch
Implementation of Graph Transformer in Pytorch, for potential use in replicating Alphafold2
Stars: ✭ 81 (-36.72%)
Mutual labels:  transformer
GTSRB Keras STN
German Traffic Sign Recognition Benchmark, Keras implementation with Spatial Transformer Networks
Stars: ✭ 48 (-62.5%)
Mutual labels:  transformer
segmenter
[ICCV2021] Official PyTorch implementation of Segmenter: Transformer for Semantic Segmentation
Stars: ✭ 463 (+261.72%)
Mutual labels:  transformer
61-120 of 540 similar projects