All Projects → sarnikowski → danish_transformers

sarnikowski / danish_transformers

Licence: CC-BY-4.0 license
A collection of Danish Transformers

Projects that are alternatives of or similar to danish transformers

converse
Conversational text Analysis using various NLP techniques
Stars: ✭ 147 (+390%)
Mutual labels:  transformers, huggingface
clip-italian
CLIP (Contrastive Language–Image Pre-training) for Italian
Stars: ✭ 113 (+276.67%)
Mutual labels:  transformers, huggingface
HugsVision
HugsVision is a easy to use huggingface wrapper for state-of-the-art computer vision
Stars: ✭ 154 (+413.33%)
Mutual labels:  transformers, huggingface
awesome-huggingface
🤗 A list of wonderful open-source projects & applications integrated with Hugging Face libraries.
Stars: ✭ 436 (+1353.33%)
Mutual labels:  transformers, huggingface
chef-transformer
Chef Transformer 🍲 .
Stars: ✭ 29 (-3.33%)
Mutual labels:  transformers, huggingface
erc
Emotion recognition in conversation
Stars: ✭ 34 (+13.33%)
Mutual labels:  transformers, huggingface
policy-data-analyzer
Building a model to recognize incentives for landscape restoration in environmental policies from Latin America, the US and India. Bringing NLP to the world of policy analysis through an extensible framework that includes scraping, preprocessing, active learning and text analysis pipelines.
Stars: ✭ 22 (-26.67%)
Mutual labels:  transformers, huggingface
naru
Neural Relation Understanding: neural cardinality estimators for tabular data
Stars: ✭ 76 (+153.33%)
Mutual labels:  transformers
Transformer-in-PyTorch
Transformer/Transformer-XL/R-Transformer examples and explanations
Stars: ✭ 21 (-30%)
Mutual labels:  transformers
LIT
[AAAI 2022] This is the official PyTorch implementation of "Less is More: Pay Less Attention in Vision Transformers"
Stars: ✭ 79 (+163.33%)
Mutual labels:  transformers
SnowflakeNet
(TPAMI 2022) Snowflake Point Deconvolution for Point Cloud Completion and Generation with Skip-Transformer
Stars: ✭ 74 (+146.67%)
Mutual labels:  transformers
hugdatafast
The elegant integration of huggingface/nlp and fastai2 and handy transforms using pure huggingface/nlp
Stars: ✭ 19 (-36.67%)
Mutual labels:  huggingface
AnimeGANv3
Use AnimeGANv3 to make your own animation works, including turning photos or videos into anime.
Stars: ✭ 878 (+2826.67%)
Mutual labels:  huggingface
KB-ALBERT
KB국민은행에서 제공하는 경제/금융 도메인에 특화된 한국어 ALBERT 모델
Stars: ✭ 215 (+616.67%)
Mutual labels:  transformers
uniformer-pytorch
Implementation of Uniformer, a simple attention and 3d convolutional net that achieved SOTA in a number of video classification tasks, debuted in ICLR 2022
Stars: ✭ 90 (+200%)
Mutual labels:  transformers
Basic-UI-for-GPT-J-6B-with-low-vram
A repository to run gpt-j-6b on low vram machines (4.2 gb minimum vram for 2000 token context, 3.5 gb for 1000 token context). Model loading takes 12gb free ram.
Stars: ✭ 90 (+200%)
Mutual labels:  transformers
jax-models
Unofficial JAX implementations of deep learning research papers
Stars: ✭ 108 (+260%)
Mutual labels:  transformers
thermostat
Collection of NLP model explanations and accompanying analysis tools
Stars: ✭ 126 (+320%)
Mutual labels:  transformers
Transformer Temporal Tagger
Code and data form the paper BERT Got a Date: Introducing Transformers to Temporal Tagging
Stars: ✭ 55 (+83.33%)
Mutual labels:  huggingface
TabFormer
Code & Data for "Tabular Transformers for Modeling Multivariate Time Series" (ICASSP, 2021)
Stars: ✭ 209 (+596.67%)
Mutual labels:  huggingface

🤖 Danish Transformers

Transformers constitute the current paradigm within Natural Language Processing (NLP) for a variety of downstream tasks. The number of transformers trained on danish corpora are limited, which is why the ambition of this repository is to provide the danish NLP community with alternatives to already established models. The pretrained models in this repository are trained using 🤗Transformers, and checkpoints are made available at the HuggingFace model hub here for both PyTorch and TensorFlow.

Model Weights

Details on how to use the models, can be found by clicking the architecture headers.

ConvBERT

ELECTRA

**Pretrained using ELECTRA pretraining approach.

Benchmarks

All downstream task benchmarks are evaluated on finetuned versions of the transformer models. The dataset used for benchmarking both NER and POS tagging, is the Danish Dependency Treebank UD-DDT. All models were trained for 3 epochs on the train set. All scores reported, are averages calculated from (N=5) random seed runs for each model, where σ refers to the standard deviation.

Named Entity Recognition

The table below shows the F1-scores on the test+dev set on the entities LOC, ORG, PER and MISC over (N=5) runs.

Model Params LOC ORG PER MISC Micro AVG
bert-base-multilingual-cased ~177M 87.02 75.24 91.28 75.94 83.18 (σ=0.81)
danish-bert-uncased-v2 ~110M 87.40 75.43 93.92 76.21 84.19 (σ=0.75)
+++++++++++++++++++++++++++ +++++ ++++ ++++ ++++ ++++ +++++++++++
convbert-medium-small-da-cased ~24.3M 88.61 75.97 90.15 77.07 83.54 (σ=0.55)
convbert-small-da-cased ~12.9M 85.86 71.21 89.07 73.50 80.76 (σ=0.40)
electra-small-da-cased ~13.3M 86.30 70.05 88.34 71.31 79.63 (σ=0.22)

Part-of-speech Tagging

The table below shows the F1-scores on the test+dev set over (N=5) runs.

Model Params Micro AVG
bert-base-multilingual-cased ~177M 97.42 (σ=0.09)
danish-bert-uncased-v2 ~110M 98.08 (σ=0.05)
+++++++++++++++++++++++++++ +++++ +++++++++++
convbert-medium-small-da-cased ~24.3M 97.92 (σ=0.03)
convbert-small-da-cased ~12.9M 97.32 (σ=0.03)
electra-small-da-cased ~13.3M 97.42 (σ=0.05)

Data

The custom danish corpora used for pretraining, was created from the following sources:

All characters in the corpus were transliterated to ASCII with the exception of æøåÆØŧ. Sources containing web crawled data, were cleaned of overrepresented NSFW ads and commercials. The final dataset consists of 14,483,456 precomputed tensors of length 256.

References

Cite this work

to cite this work please use

@inproceedings{danish-transformers,
  title = {Danish Transformers},
  author = {Tamimi-Sarnikowski, Philip},
  year = {2020},
  publisher = {{GitHub}},
  url = {https://github.com/sarnikowski}
}

License

CC BY 4.0

This work is licensed under a Creative Commons Attribution 4.0 International License.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].