All Projects → explosion → Spacy Transformers

explosion / Spacy Transformers

Licence: mit
🛸 Use pretrained transformers like BERT, XLNet and GPT-2 in spaCy

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Spacy Transformers

Transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Stars: ✭ 55,742 (+5965.51%)
Mutual labels:  natural-language-processing, language-model, natural-language-understanding
Easy Bert
A Dead Simple BERT API for Python and Java (https://github.com/google-research/bert)
Stars: ✭ 106 (-88.47%)
Mutual labels:  natural-language-processing, language-model, natural-language-understanding
Tokenizers
💥 Fast State-of-the-Art Tokenizers optimized for Research and Production
Stars: ✭ 5,077 (+452.45%)
Mutual labels:  natural-language-processing, language-model, natural-language-understanding
Transfer Nlp
NLP library designed for reproducible experimentation management
Stars: ✭ 287 (-68.77%)
Mutual labels:  transfer-learning, language-model, natural-language-understanding
Attention Mechanisms
Implementations for a family of attention mechanisms, suitable for all kinds of natural language processing tasks and compatible with TensorFlow 2.0 and Keras.
Stars: ✭ 203 (-77.91%)
Mutual labels:  natural-language-processing, language-model, natural-language-understanding
Chars2vec
Character-based word embeddings model based on RNN for handling real world texts
Stars: ✭ 130 (-85.85%)
Mutual labels:  natural-language-processing, language-model, natural-language-understanding
Awesome Bert Nlp
A curated list of NLP resources focused on BERT, attention mechanism, Transformer networks, and transfer learning.
Stars: ✭ 567 (-38.3%)
Mutual labels:  natural-language-processing, transfer-learning, language-model
Bert Sklearn
a sklearn wrapper for Google's BERT model
Stars: ✭ 182 (-80.2%)
Mutual labels:  natural-language-processing, transfer-learning, language-model
Turkish Morphology
A two-level morphological analyzer for Turkish.
Stars: ✭ 121 (-86.83%)
Mutual labels:  google, natural-language-processing, natural-language-understanding
Bert
TensorFlow code and pre-trained models for BERT
Stars: ✭ 29,971 (+3161.26%)
Mutual labels:  google, natural-language-processing, natural-language-understanding
Practical Nlp
Official Repository for 'Practical Natural Language Processing' by O'Reilly Media
Stars: ✭ 452 (-50.82%)
Mutual labels:  natural-language-processing, natural-language-understanding
Spacy
💫 Industrial-strength Natural Language Processing (NLP) in Python
Stars: ✭ 21,978 (+2291.51%)
Mutual labels:  natural-language-processing, spacy
Nlp Paper
NLP Paper
Stars: ✭ 484 (-47.33%)
Mutual labels:  transfer-learning, language-model
Spacy Stanza
💥 Use the latest Stanza (StanfordNLP) research models directly in spaCy
Stars: ✭ 508 (-44.72%)
Mutual labels:  natural-language-processing, spacy
Seqgan
A simplified PyTorch implementation of "SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient." (Yu, Lantao, et al.)
Stars: ✭ 502 (-45.38%)
Mutual labels:  natural-language-processing, natural-language-understanding
Botlibre
An open platform for artificial intelligence, chat bots, virtual agents, social media automation, and live chat automation.
Stars: ✭ 412 (-55.17%)
Mutual labels:  natural-language-processing, natural-language-understanding
Deberta
The implementation of DeBERTa
Stars: ✭ 541 (-41.13%)
Mutual labels:  language-model, natural-language-understanding
Projects
🪐 End-to-end NLP workflows from prototype to production
Stars: ✭ 397 (-56.8%)
Mutual labels:  natural-language-processing, spacy
Chat
基于自然语言理解与机器学习的聊天机器人,支持多用户并发及自定义多轮对话
Stars: ✭ 516 (-43.85%)
Mutual labels:  natural-language-processing, natural-language-understanding
Speech Emotion Analyzer
The neural network model is capable of detecting five different male/female emotions from audio speeches. (Deep Learning, NLP, Python)
Stars: ✭ 633 (-31.12%)
Mutual labels:  natural-language-processing, natural-language-understanding

spacy-transformers: Use pretrained transformers like BERT, XLNet and GPT-2 in spaCy

This package provides spaCy components and architectures to use transformer models via Hugging Face's transformers in spaCy. The result is convenient access to state-of-the-art transformer architectures, such as BERT, GPT-2, XLNet, etc.

This release requires spaCy v3. For the previous version of this library, see the v0.6.x branch.

Azure Pipelines PyPi GitHub Code style: black

Features

  • Use pretrained transformer models like BERT, RoBERTa and XLNet to power your spaCy pipeline.
  • Easy multi-task learning: backprop to one transformer model from several pipeline components.
  • Train using spaCy v3's powerful and extensible config system.
  • Automatic alignment of transformer output to spaCy's tokenization.
  • Easily customize what transformer data is saved in the Doc object.
  • Easily customize how long documents are processed.
  • Out-of-the-box serialization and model packaging.

🚀 Installation

Installing the package from pip will automatically install all dependencies, including PyTorch and spaCy. Make sure you install this package before you install the models. Also note that this package requires Python 3.6+, PyTorch v1.5+ and spaCy v3.0+.

pip install spacy[transformers]

For GPU installation, find your CUDA version using nvcc --version and add the version in brackets, e.g. spacy[transformers,cuda92] for CUDA9.2 or spacy[transformers,cuda100] for CUDA10.0.

If you are having trouble installing PyTorch, follow the instructions on the official website for your specific operation system and requirements, or try the following:

pip install spacy-transformers -f https://download.pytorch.org/whl/torch_stable.html

📖 Documentation

⚠️ Important note: This package has been extensively refactored to take advantage of spaCy v3.0. Previous versions that were built for spaCy v2.x worked considerably differently. Please see previous tagged versions of this README for documentation on prior versions.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].