knodleA PyTorch-based open-source framework that provides methods for improving the weakly annotated data and allows researchers to efficiently develop and compare their own methods.
Stars: ✭ 76 (-58.92%)
WeFEND-AAAI20Dataset for paper "Weak Supervision for Fake News Detection via Reinforcement Learning" published in AAAI'2020.
Stars: ✭ 67 (-63.78%)
troveWeakly supervised medical named entity classification
Stars: ✭ 55 (-70.27%)
concept-based-xaiLibrary implementing state-of-the-art Concept-based and Disentanglement Learning methods for Explainable AI
Stars: ✭ 41 (-77.84%)
Learning-From-RulesImplementation of experiments in paper "Learning from Rules Generalizing Labeled Exemplars" to appear in ICLR2020 (https://openreview.net/forum?id=SkeuexBtDr)
Stars: ✭ 46 (-75.14%)
spearSPEAR: Programmatically label and build training data quickly.
Stars: ✭ 81 (-56.22%)
ASTRASelf-training with Weak Supervision (NAACL 2021)
Stars: ✭ 127 (-31.35%)
weaselWeakly Supervised End-to-End Learning (NeurIPS 2021)
Stars: ✭ 117 (-36.76%)
Lstm CrfA (CNN+)RNN(LSTM/BiLSTM)+CRF model for sequence labelling.😏
Stars: ✭ 134 (-27.57%)
ProsodyHelsinki Prosody Corpus and A System for Predicting Prosodic Prominence from Text
Stars: ✭ 139 (-24.86%)
NcrfppNCRF++, a Neural Sequence Labeling Toolkit. Easy use to any sequence labeling tasks (e.g. NER, POS, Segmentation). It includes character LSTM/CNN, word LSTM/CNN and softmax/CRF components.
Stars: ✭ 1,767 (+855.14%)
weak-supervision-for-NERFramework to learn Named Entity Recognition models without labelled data using weak supervision.
Stars: ✭ 114 (-38.38%)
FlairA very simple framework for state-of-the-art Natural Language Processing (NLP)
Stars: ✭ 11,065 (+5881.08%)
emp🔬 Empirical CLI
Stars: ✭ 42 (-77.3%)
LightnerInference with state-of-the-art models (pre-trained by LD-Net / AutoNER / VanillaNER / ...)
Stars: ✭ 102 (-44.86%)
Ntaggerreference pytorch code for named entity tagging
Stars: ✭ 58 (-68.65%)
WearableSensorDataThis repository provides the codes and data used in our paper "Human Activity Recognition Based on Wearable Sensor Data: A Standardization of the State-of-the-Art", where we implement and evaluate several state-of-the-art approaches, ranging from handcrafted-based methods to convolutional neural networks.
Stars: ✭ 65 (-64.86%)
SciteCausality Extraction based on Self-Attentive BiLSTM-CRF with Transferred Embeddings
Stars: ✭ 22 (-88.11%)
Cluener2020CLUENER2020 中文细粒度命名实体识别 Fine Grained Named Entity Recognition
Stars: ✭ 689 (+272.43%)
CleanlabThe standard package for machine learning with noisy labels, finding mislabeled data, and uncertainty quantification. Works with most datasets and models.
Stars: ✭ 2,526 (+1265.41%)
Sequence Labeling Bilstm CrfThe classical BiLSTM-CRF model implemented in Tensorflow, for sequence labeling tasks. In Vex version, everything is configurable.
Stars: ✭ 579 (+212.97%)
SeqevalA Python framework for sequence labeling evaluation(named-entity recognition, pos tagging, etc...)
Stars: ✭ 508 (+174.59%)
Neuronlp2Deep neural models for core NLP tasks (Pytorch version)
Stars: ✭ 397 (+114.59%)
Ld NetEfficient Contextualized Representation: Language Model Pruning for Sequence Labeling
Stars: ✭ 148 (-20%)
RHEL8-STIGAnsible role for Red Hat 8 STIG Baseline
Stars: ✭ 73 (-60.54%)
Nlp Projectsword2vec, sentence2vec, machine reading comprehension, dialog system, text classification, pretrained language model (i.e., XLNet, BERT, ELMo, GPT), sequence labeling, information retrieval, information extraction (i.e., entity, relation and event extraction), knowledge graph, text generation, network embedding
Stars: ✭ 360 (+94.59%)
Sltk序列化标注工具,基于PyTorch实现BLSTM-CNN-CRF模型,CoNLL 2003 English NER测试集F1值为91.10%(word and char feature)。
Stars: ✭ 338 (+82.7%)
Aspect ExtractionAspect extraction from product reviews - window-CNN+maxpool+CRF, BiLSTM+CRF, MLP+CRF
Stars: ✭ 117 (-36.76%)
cbtoolCloud Rapid Experimentation and Analysis Toolkit
Stars: ✭ 65 (-64.86%)
AnagoBidirectional LSTM-CRF and ELMo for Named-Entity Recognition, Part-of-Speech Tagging and so on.
Stars: ✭ 1,392 (+652.43%)
UBUNTU20-CISAnsible role for Ubuntu 2004 CIS Baseline
Stars: ✭ 136 (-26.49%)
NeuronblocksNLP DNN Toolkit - Building Your NLP DNN Models Like Playing Lego
Stars: ✭ 1,356 (+632.97%)
rwloadsimRWP*Load Simulator - your tool for scripting, simulation and much more. Like having a bit of bash and SQL, a nip of C or Java, a dash of awk, a grain of sed plus drops of secret sauce in one single tool. See https://blogs.oracle.com/database/rwploadsim-oracle-db-performance-simluator for the announcement on the Oracle database blog.
Stars: ✭ 26 (-85.95%)
hamnetPyTorch implementation of AAAI 2021 paper: A Hybrid Attention Mechanism for Weakly-Supervised Temporal Action Localization
Stars: ✭ 30 (-83.78%)
Bilstm LanHierarchically-Refined Label Attention Network for Sequence Labeling
Stars: ✭ 241 (+30.27%)
Bert For Sequence Labeling And Text ClassificationThis is the template code to use BERT for sequence lableing and text classification, in order to facilitate BERT for more tasks. Currently, the template code has included conll-2003 named entity identification, Snips Slot Filling and Intent Prediction.
Stars: ✭ 293 (+58.38%)
Lm Lstm CrfEmpower Sequence Labeling with Task-Aware Language Model
Stars: ✭ 778 (+320.54%)
cpmContinuous Perfomance Monitor (CPM) for C++ code
Stars: ✭ 39 (-78.92%)
MCIS wsssCode for ECCV 2020 paper (oral): Mining Cross-Image Semantics for Weakly Supervised Semantic Segmentation
Stars: ✭ 151 (-18.38%)
Rnn NluA TensorFlow implementation of Recurrent Neural Networks for Sequence Classification and Sequence Labeling
Stars: ✭ 463 (+150.27%)
uber-graph-benchmarkA framework to benchmark different graph databases, based on generated data from customizable schema, distribution, and size.
Stars: ✭ 25 (-86.49%)
AutonerLearning Named Entity Tagger from Domain-Specific Dictionary
Stars: ✭ 357 (+92.97%)
Pytorch ner bilstm cnn crfEnd-to-end Sequence Labeling via Bi-directional LSTM-CNNs-CRF implement in pyotrch
Stars: ✭ 249 (+34.59%)
Slot filling and intent detection of sluslot filling, intent detection, joint training, ATIS & SNIPS datasets, the Facebook’s multilingual dataset, MIT corpus, E-commerce Shopping Assistant (ECSA) dataset, CoNLL2003 NER, ELMo, BERT, XLNet
Stars: ✭ 298 (+61.08%)
SnorkelA system for quickly generating training data with weak supervision
Stars: ✭ 4,953 (+2577.3%)
Multi Task Nlpmulti_task_NLP is a utility toolkit enabling NLP developers to easily train and infer a single model for multiple tasks.
Stars: ✭ 221 (+19.46%)
Delfta Deep Learning Framework for Text
Stars: ✭ 289 (+56.22%)
GectorOfficial implementation of the paper “GECToR – Grammatical Error Correction: Tag, Not Rewrite” // Published on BEA15 Workshop (co-located with ACL 2020) https://www.aclweb.org/anthology/2020.bea-1.16.pdf
Stars: ✭ 287 (+55.14%)
Hscrf PytorchACL 2018: Hybrid semi-Markov CRF for Neural Sequence Labeling (http://aclweb.org/anthology/P18-2038)
Stars: ✭ 284 (+53.51%)
Neural sequence labelingA TensorFlow implementation of Neural Sequence Labeling model, which is able to tackle sequence labeling tasks such as POS Tagging, Chunking, NER, Punctuation Restoration and etc.
Stars: ✭ 214 (+15.68%)
Rnn For Joint NluTensorflow implementation of "Attention-Based Recurrent Neural Network Models for Joint Intent Detection and Slot Filling" (https://arxiv.org/abs/1609.01454)
Stars: ✭ 281 (+51.89%)
reefAutomatically labeling training data
Stars: ✭ 102 (-44.86%)
pyner🌈 Implementation of Neural Network based Named Entity Recognizer (Lample+, 2016) using Chainer.
Stars: ✭ 45 (-75.68%)
KashgariKashgari is a production-level NLP Transfer learning framework built on top of tf.keras for text-labeling and text-classification, includes Word2Vec, BERT, and GPT2 Language Embedding.
Stars: ✭ 2,235 (+1108.11%)
RnnsharpRNNSharp is a toolkit of deep recurrent neural network which is widely used for many different kinds of tasks, such as sequence labeling, sequence-to-sequence and so on. It's written by C# language and based on .NET framework 4.6 or above versions. RNNSharp supports many different types of networks, such as forward and bi-directional network, sequence-to-sequence network, and different types of layers, such as LSTM, Softmax, sampled Softmax and others.
Stars: ✭ 277 (+49.73%)