All Projects → xuuuluuu → SynLSTM-for-NER

xuuuluuu / SynLSTM-for-NER

Licence: other
Code and models for the paper titled "Better Feature Integration for Named Entity Recognition", NAACL 2021.

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to SynLSTM-for-NER

Bertner
ChineseNER based on BERT, with BiLSTM+CRF layer
Stars: ✭ 195 (+650%)
Mutual labels:  named-entity-recognition, ner
neural name tagging
Code for "Reliability-aware Dynamic Feature Composition for Name Tagging" (ACL2019)
Stars: ✭ 39 (+50%)
Mutual labels:  named-entity-recognition, ner
Monpa
MONPA 罔拍是一個提供正體中文斷詞、詞性標註以及命名實體辨識的多任務模型
Stars: ✭ 203 (+680.77%)
Mutual labels:  named-entity-recognition, ner
Kashgari
Kashgari is a production-level NLP Transfer learning framework built on top of tf.keras for text-labeling and text-classification, includes Word2Vec, BERT, and GPT2 Language Embedding.
Stars: ✭ 2,235 (+8496.15%)
Mutual labels:  named-entity-recognition, ner
Bert ner
Ner with Bert
Stars: ✭ 240 (+823.08%)
Mutual labels:  named-entity-recognition, ner
Bert Sklearn
a sklearn wrapper for Google's BERT model
Stars: ✭ 182 (+600%)
Mutual labels:  named-entity-recognition, ner
NER-and-Linking-of-Ancient-and-Historic-Places
An NER tool for ancient place names based on Pleiades and Spacy.
Stars: ✭ 26 (+0%)
Mutual labels:  named-entity-recognition, ner
Bnlp
BNLP is a natural language processing toolkit for Bengali Language.
Stars: ✭ 127 (+388.46%)
Mutual labels:  named-entity-recognition, ner
PhoNER COVID19
COVID-19 Named Entity Recognition for Vietnamese (NAACL 2021)
Stars: ✭ 55 (+111.54%)
Mutual labels:  named-entity-recognition, ner
Pytorch Bert Crf Ner
KoBERT와 CRF로 만든 한국어 개체명인식기 (BERT+CRF based Named Entity Recognition model for Korean)
Stars: ✭ 236 (+807.69%)
Mutual labels:  named-entity-recognition, ner
Sequence tagging
Named Entity Recognition (LSTM + CRF) - Tensorflow
Stars: ✭ 1,889 (+7165.38%)
Mutual labels:  named-entity-recognition, ner
KoBERT-NER
NER Task with KoBERT (with Naver NLP Challenge dataset)
Stars: ✭ 76 (+192.31%)
Mutual labels:  named-entity-recognition, ner
Ld Net
Efficient Contextualized Representation: Language Model Pruning for Sequence Labeling
Stars: ✭ 148 (+469.23%)
Mutual labels:  named-entity-recognition, ner
Persian Ner
پیکره بزرگ شناسایی موجودیت‌های نامدار فارسی برچسب خورده
Stars: ✭ 183 (+603.85%)
Mutual labels:  named-entity-recognition, ner
Ncrfpp
NCRF++, a Neural Sequence Labeling Toolkit. Easy use to any sequence labeling tasks (e.g. NER, POS, Segmentation). It includes character LSTM/CNN, word LSTM/CNN and softmax/CRF components.
Stars: ✭ 1,767 (+6696.15%)
Mutual labels:  named-entity-recognition, ner
Spacy Lookup
Named Entity Recognition based on dictionaries
Stars: ✭ 212 (+715.38%)
Mutual labels:  named-entity-recognition, ner
Dan Jurafsky Chris Manning Nlp
My solution to the Natural Language Processing course made by Dan Jurafsky, Chris Manning in Winter 2012.
Stars: ✭ 124 (+376.92%)
Mutual labels:  named-entity-recognition, ner
Ner Evaluation
An implementation of a full named-entity evaluation metrics based on SemEval'13 Task 9 - not at tag/token level but considering all the tokens that are part of the named-entity
Stars: ✭ 126 (+384.62%)
Mutual labels:  named-entity-recognition, ner
Ner Datasets
Datasets to train supervised classifiers for Named-Entity Recognition in different languages (Portuguese, German, Dutch, French, English)
Stars: ✭ 220 (+746.15%)
Mutual labels:  named-entity-recognition, ner
Ner Bert Pytorch
PyTorch solution of named entity recognition task Using Google AI's pre-trained BERT model.
Stars: ✭ 249 (+857.69%)
Mutual labels:  named-entity-recognition, ner

Better Feature Integration for Named Entity Recognition

[NAACL 2021] Better Feature Integration for Named Entity Recognition (In NAACL 2021)

Requirement

Python 3.7

Pytorch 1.4.0

Transformers 3.3.1

CUDA 10.1, 10.2

Bert-as-service

Running

Firstly, download the embedding files: glove.6B.100d.txt , cc.ca.300.vec, cc.es.300.vec, cc.zh.300.vec, and put the files in the data folder.

By default, the model eval our saved model (without BERT) on SemEval 2010 Task 1 Spanish dataset.

python main.py  

To train the model with other datasets:

python main.py --mode=train --dataset=ontonotes --embedding_file=glove.6B.100d.txt

To train with BERT, first obtain the contextual embedding with the instructions in the get_context_emb folder (The contextual embedding files for OntoNotes Engligh can be downloaded from here.), and then run with the command:

python main.py --mode=train --dataset=ontonotes --embedding_file=glove.6B.100d.txt --context_emb=bert 

Note that the flag --dep_model=dggcn (by default) is where we call both GCN and our Syn-LSTM model. The flag --num_lstm-layer is designed for running some baselines, and should be set to 0 (by default) when running our proposed model.

About Dataset

Note that we use the data from 4 columns: word, dependency head index, dependency relation label, and entity label.

Related Repo

The code are created based on the code of the paper "Dependency-Guided LSTM-CRF Model for Named Entity Recognition", EMNLP 2019.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].