All Projects → alibaba-research → Chineseblue

alibaba-research / Chineseblue

Licence: apache-2.0
Chinese Biomedical Language Understanding Evaluation benchmark (ChineseBLUE)

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Chineseblue

Uer Py
Open Source Pre-training Model Framework in PyTorch & Pre-trained Model Zoo
Stars: ✭ 1,295 (+769.13%)
Mutual labels:  chinese, natural-language-processing
CBLUE
中文医疗信息处理基准CBLUE: A Chinese Biomedical Language Understanding Evaluation Benchmark
Stars: ✭ 379 (+154.36%)
Mutual labels:  benchmark, chinese
Medical Books
Open sourece medical books in LaTeX. LaTeX写的中文开源医学书籍
Stars: ✭ 211 (+41.61%)
Mutual labels:  chinese, medical
Medmnist
[ISBI'21] MedMNIST Classification Decathlon: A Lightweight AutoML Benchmark for Medical Image Analysis
Stars: ✭ 338 (+126.85%)
Mutual labels:  medical, benchmark
D2l Zh
《动手学深度学习》:面向中文读者、能运行、可讨论。中英文版被55个国家的300所大学用于教学。
Stars: ✭ 29,132 (+19451.68%)
Mutual labels:  chinese, natural-language-processing
Jionlp
中文 NLP 任务预处理工具包,准确、高效、零使用门槛
Stars: ✭ 449 (+201.34%)
Mutual labels:  chinese, natural-language-processing
FewCLUE
FewCLUE 小样本学习测评基准,中文版
Stars: ✭ 251 (+68.46%)
Mutual labels:  benchmark, chinese
Biomedicus
Code for the old version of BioMedICUS, for the new version see the biomedicus3 repository.
Stars: ✭ 45 (-69.8%)
Mutual labels:  natural-language-processing, medical
Clue
中文语言理解测评基准 Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained models, corpus and leaderboard
Stars: ✭ 2,425 (+1527.52%)
Mutual labels:  chinese, benchmark
Awesome Http Benchmark
HTTP(S) benchmark tools, testing/debugging, & restAPI (RESTful)
Stars: ✭ 2,236 (+1400.67%)
Mutual labels:  benchmark
Tree Transformer
Implementation of the paper Tree Transformer
Stars: ✭ 148 (-0.67%)
Mutual labels:  natural-language-processing
Leveldb Handbook
Analysis leveldb source code step by step
Stars: ✭ 146 (-2.01%)
Mutual labels:  chinese
Swiftui Tutorials
A code example and translation project of SwiftUI. / 一个 SwiftUI 的示例、翻译的教程项目。
Stars: ✭ 1,992 (+1236.91%)
Mutual labels:  chinese
Negapoji
Japanese negative positive classification.日本語文書のネガポジを判定。
Stars: ✭ 148 (-0.67%)
Mutual labels:  natural-language-processing
Metabench
A simple framework for compile-time benchmarks
Stars: ✭ 146 (-2.01%)
Mutual labels:  benchmark
Spacy Course
👩‍🏫 Advanced NLP with spaCy: A free online course
Stars: ✭ 1,920 (+1188.59%)
Mutual labels:  natural-language-processing
Hands On Natural Language Processing With Python
This repository is for my students of Udemy. You can find all lecture codes along with mentioned files for reading in here. So, feel free to clone it and if you have any problem just raise a question.
Stars: ✭ 146 (-2.01%)
Mutual labels:  natural-language-processing
Hanzi chaizi
汉字拆字库,可以将汉字拆解成偏旁部首,在机器学习中作为汉字的字形特征
Stars: ✭ 146 (-2.01%)
Mutual labels:  chinese
Spacymoji
💙 Emoji handling and meta data for spaCy with custom extension attributes
Stars: ✭ 151 (+1.34%)
Mutual labels:  natural-language-processing
Swiftychrono
A natural language date parser in Swift (ported from chrono.js)
Stars: ✭ 148 (-0.67%)
Mutual labels:  natural-language-processing

ChineseBLUE, the Chinese Biomedical Language Understanding Evaluation benchmark

License GitHub stars GitHub forks

Introduction

ChinesseBLUE benchmark consists of different biomedicine text-mining tasks with corpora. These tasks cover a diverse range of text genres (biomedical web data and clinical notes), dataset sizes, and degrees of difficulty and, more importantly, highlight common biomedicine text-mining challenges.

MC-BERT is a novel conceptualized representation learning approach for the medical domain. First, we use a different mask generation procedure to mask spans of tokens, rather than only random ones. We also introduce two kinds of masking strategies, namely whole entity masking and whole span masking. Finally, MC-BERT split the input document into segments based on the actual "sentences" provided by the user as positive samples and sample random sentences from other documents as negative samples for the next sentence prediction.

c-bert model

Tasks

Dataset Train Dev Test Task Metrics Domain
cEHRNER 914 44 41 Name Entity Recognition F1 Clinical
cMedQANER 1,673 175 215 Name Entity Recognition F1 Medical
cMedQQ 16,071 1,793 1,935 Paraphrase Identification F1 Medical
cMedQNLI 80,950 9,065 9,969 Question Natural Language Inference F1 Medical
cMedQA 49,719 5,475 6,149 Question Answering F1 Medical
cMedIR 80,000 10,000 10,000 Information Rerival MRR Medical
cMedIC 1,683 123 84 Intent Classification F1 Medical
cMedTC 1,683 123 84 Sentence Classification F1 Medical
CMDD 1240 412 412 Symptom Diagnosis F1 Dialogue

Named Entity Recognition (NER)

Named entity recognition aims to recognize various entities, including diseases, drugs, syndromes, etc. The cEHRNER dataset labeled from the Chinese electronic health records and the cMedQANER dataset labeled from Chinese community question answering is chosen.

Paraphrase Identification (PI)

Paraphrase Identification aims to identify whether two sentences express the same meaning. We use cMedQQ, which consists of search query pairs.

Question Natural Language Inference (QNLI)

Question natural language inference aims to identify whether the answer corresponds to the question in the question-answering pair. We use cMedQNLI, which consists of question-answer pairs.

Question Answering (QA)

Question answering can be approximated as ranking candidate answer sentences based on their similarity. We assign 0,1 labels to the QA pairs, which convert to the binary classification problem. We use cMedQA released from the paper "Multi-Scale Attentive Interaction Networks for Chinese Medical Question Answer Selection", which consists of questions and their answers.

Information Retrieval (IR)

Information retrieval aims to retrieve most related documents given search queries. IR can be regarded as a ranking task. We use the cMedIR dataset, which consists of queries with multiple documents and their relative scores.

Intent Classification (IC)

Intent classification aims to assign intent labels to the queries, which can be regarded as multiple label classification tasks. We use the cMedIC dataset, which consists of queries with three intent labels (e.g., no intention, weak intention, and firm intention).

Text Classification (TC)

Text classification aims to assign multiple labels to the sentence. We use the cMedTC dataset, which consists of biomedical texts with multiple labels.

Symptom Diagnosis

Symptom diagnosis is a challenging yet profound problem in natural language processing. We use the CMDD dataset released from the paper"Enhancing Dialogue Symptom Diagnosis with Global Attention and Symptom Graph".

Relation extraction (RE)

Coming soon.

Datasets

All datasets can be downloaded at ChineseBLUE1.0

Pretrained Model

All pre-trained models can be downloaded at MC-BERT.

Citing Chinese BLUE

Paper will be released soon.

@article{zhang2020conceptualized,
  title={Conceptualized Representation Learning for Chinese Biomedical Text Mining},
  author={Zhang, Ningyu and Jia, Qianghuai and Yin, Kangping and Dong, Liang and Gao, Feng and Hua, Nengwei},
  journal={arXiv preprint arXiv:2008.10813},
  year={2020}
}

Acknowledgments

We are also grateful to the authors of BERT and wwm-BERT to make the data and codes publicly available. We are also grateful to the authors of paper "Enhancing Dialogue Symptom Diagnosis with Global Attention and Symptom Graph" and "Multi-Scale Attentive Interaction Networks for Chinese Medical Question Answer Selection" who released the dataset CMDD and cMedQA2.

Disclaimer

This project is not the official product of Alibaba. The information produced on this website is not intended for direct diagnostic use or medical decision-making without review and oversight by a clinical professional. Individuals should not change their health behavior solely on the basis of information produced on this website. If you have questions about the information produced on this website, please see a health care professional.

The experimental results presented in the technical report only indicate the performance under the specific data set and super-parameter combination and do not represent the essence of each model. The experimental results may change due to random number seeds and computing devices. The content of this project is for technical research purposes only and is not intended to be a conclusive basis. The user may use the model arbitrarily within the scope of the license, but we are not responsible for any direct or indirect damages resulting from the use of the content of the project.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].