All Projects → jiachenwestlake → Cross Domain_ner

jiachenwestlake / Cross Domain_ner

Cross-domain NER using cross-domain language modeling, code for ACL 2019 paper

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Cross Domain ner

cmd
Central Moment Discrepancy for Domain-Invariant Representation Learning (ICLR 2017, keras)
Stars: ✭ 53 (-20.9%)
Mutual labels:  transfer-learning, domain-adaptation
Transfer Nlp
NLP library designed for reproducible experimentation management
Stars: ✭ 287 (+328.36%)
Mutual labels:  transfer-learning, language-model
KD3A
Here is the official implementation of the model KD3A in paper "KD3A: Unsupervised Multi-Source Decentralized Domain Adaptation via Knowledge Distillation".
Stars: ✭ 63 (-5.97%)
Mutual labels:  transfer-learning, domain-adaptation
meta-learning-progress
Repository to track the progress in Meta-Learning (MtL), including the datasets and the current state-of-the-art for the most common MtL problems.
Stars: ✭ 26 (-61.19%)
Mutual labels:  transfer-learning, domain-adaptation
Transfer Learning Library
Transfer-Learning-Library
Stars: ✭ 678 (+911.94%)
Mutual labels:  transfer-learning, domain-adaptation
transfer-learning-algorithms
Implementation of many transfer learning algorithms in Python with Jupyter notebooks
Stars: ✭ 42 (-37.31%)
Mutual labels:  transfer-learning, domain-adaptation
adapt
Awesome Domain Adaptation Python Toolbox
Stars: ✭ 46 (-31.34%)
Mutual labels:  transfer-learning, domain-adaptation
BA3US
code for our ECCV 2020 paper "A Balanced and Uncertainty-aware Approach for Partial Domain Adaptation"
Stars: ✭ 31 (-53.73%)
Mutual labels:  transfer-learning, domain-adaptation
Awesome Bert Nlp
A curated list of NLP resources focused on BERT, attention mechanism, Transformer networks, and transfer learning.
Stars: ✭ 567 (+746.27%)
Mutual labels:  transfer-learning, language-model
Nlp Paper
NLP Paper
Stars: ✭ 484 (+622.39%)
Mutual labels:  transfer-learning, language-model
wechsel
Code for WECHSEL: Effective initialization of subword embeddings for cross-lingual transfer of monolingual language models.
Stars: ✭ 39 (-41.79%)
Mutual labels:  transfer-learning, language-model
Bert language understanding
Pre-training of Deep Bidirectional Transformers for Language Understanding: pre-train TextCNN
Stars: ✭ 933 (+1292.54%)
Mutual labels:  transfer-learning, language-model
backprop
Backprop makes it simple to use, finetune, and deploy state-of-the-art ML models.
Stars: ✭ 229 (+241.79%)
Mutual labels:  transfer-learning, language-model
MinTL
MinTL: Minimalist Transfer Learning for Task-Oriented Dialogue Systems
Stars: ✭ 61 (-8.96%)
Mutual labels:  transfer-learning, language-model
Transformers-Domain-Adaptation
Adapt Transformer-based language models to new text domains
Stars: ✭ 67 (+0%)
Mutual labels:  transfer-learning, domain-adaptation
SHOT-plus
code for our TPAMI 2021 paper "Source Data-absent Unsupervised Domain Adaptation through Hypothesis Transfer and Labeling Transfer"
Stars: ✭ 46 (-31.34%)
Mutual labels:  transfer-learning, domain-adaptation
pykale
Knowledge-Aware machine LEarning (KALE): accessible machine learning from multiple sources for interdisciplinary research, part of the 🔥PyTorch ecosystem
Stars: ✭ 381 (+468.66%)
Mutual labels:  transfer-learning, domain-adaptation
Transfer-learning-materials
resource collection for transfer learning!
Stars: ✭ 213 (+217.91%)
Mutual labels:  transfer-learning, domain-adaptation
Multitask Learning
Awesome Multitask Learning Resources
Stars: ✭ 361 (+438.81%)
Mutual labels:  transfer-learning, domain-adaptation
Spacy Transformers
🛸 Use pretrained transformers like BERT, XLNet and GPT-2 in spaCy
Stars: ✭ 919 (+1271.64%)
Mutual labels:  transfer-learning, language-model

Cross-Domain_NER

Cross-domain NER using cross-domain language modeling, code for ACL 2019 paper.

Introduction

NER is a fundamental task in NLP. Due to the limitation of labeled resources, cross-domain NER has been a challenging task. Most previous work concentrates on the supervised scenario, making use of labeled data for both source and target domains. A disadvantage of such setting is that they can not train for domains which have no labeled data.

We address this issue, using cross-domain LM as a bridge cross-domains for NER domain adaptation. Performing cross-task and cross-domain transfer by designing a novel Parameter Generation Network.

Experiments on CBS SciTech News Dataset show that our model can effectively allow unsupervised domain adaptation, while also can deriving supervised domain adaption between domains with completely different entity types (i.e. news vs. biomedical).

The naive baseline of Single Task Model (STM in paper) mostly followed NCRF++.

For more details, please refer to our paper:

Cross-Domain NER using Cross-Domain Language Modeling
Chen Jia, Xiaobo Liang and Yue Zhang*
(* Corresponding Author)
ACL 2019

Requirements

Python 2 or 3 
PyTorch 0.3

The memory of one GPU should be no less than 8GB to fit the model.

Pretrained Embeddings

GloVe 100-dimension word vectors (Cite from Here).

DataSet

Source-domain:

CoNLL-2003 English NER data.

Target-domain

  • Unsupervised: CBS SciTech News (test set) (In: `\unsupervised_domain_adaptation\data\news_tech\tech_test).
  • Supervised: BioNLP13PC dataset and BioNLP13CG dataset.

LM raw data

Retures domain raw data is released together with the CoNLL-2003 dataset.
SciTech news domain raw data Download.
Reuters news domain raw data Download.

Usage

Command

\supervised_domain_adaptation, \unsupervised_domain_adaptationand \combined_SDA_and_UDA can use the following command to make it run.

python main.py --config train.NER.config

The file train.NER.config contains dataset path and model hyperparameters following NCRF++.

Input format

  • We recommand using the IBOES label style for NER dataset.
  • We recommand using an input style of one-sentence-per-line for raw data with word segmentation.

Cite:

If you use our data or code, please cite our paper as follows:

@inproceedings{jia2019cross,
  title={Cross-domain ner using cross-domain language modeling},
  author={Jia, Chen and Liang, Xiaobo and Zhang, Yue},
  booktitle={Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics},
  pages={2464--2474},
  year={2019}
  organization={Association for Computational Linguistics}
}

Update

  • V2. Combining supervised scenario and unsupervised scenario in \combined_SDA_and_UDA.
  • V1. The previous supervised scenario in \supervised_domain_adaptation;
    The previous unsupervised scenario in \unsupervised_domain_adaptation;
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].