Skip to content

Chinese Biomedical Language Understanding Evaluation benchmark (ChineseBLUE)

License

Notifications You must be signed in to change notification settings

chosendai/ChineseBLUE

Repository files navigation

ChineseBLUE, the Chinese Biomedical Language Understanding Evaluation benchmark

License GitHub stars GitHub forks HitCount

Introduction

ChinesseBLUE benchmark consists of different biomedicine text-mining tasks with corpora. These tasks cover a diverse range of text genres (biomedical web data and clinical notes), dataset sizes, and degrees of difficulty and, more importantly, highlight common biomedicine text-mining challenges.

MC-BERT is a novel conceptualized representation learning approach for the medical domain. First, we use a different mask generation procedure to mask spans of tokens, rather than only random ones. We also introduce two kinds of masking strategies, namely whole entity masking and whole span masking. Finally, MC-BERT split the input document into segments based on the actual "sentences" provided by the user as positive samples and sample random sentences from other documents as negative samples for the next sentence prediction.

c-bert model

Tasks

Dataset Train Dev Test Task Metrics Domain
cEHRNER 914 44 41 Name Entity Recognition F1 Clinical
cMedQANER 1,673 175 215 Name Entity Recognition F1 Medical
cMedQQ 16,071 1,793 1,935 Paraphrase Identification F1 Medical
cMedQNLI 80,950 9,065 9,969 Question Natural Language Inference F1 Medical
cMedQA 49,719 5,475 6,149 Question Answering F1 Medical
cMedIR 80,000 10,000 10,000 Information Rerival MRR Medical
cMedIC 1,683 123 84 Intent Classification F1 Medical
cMedTC 1,683 123 84 Sentence Classification F1 Medical
CMDD 1240 412 412 Symptom Diagnosis F1 Dialogue

Named Entity Recognition (NER)

Named entity recognition aims to recognize various entities, including diseases, drugs, syndromes, etc. The cEHRNER dataset labeled from the Chinese electronic health records and the cMedQANER dataset labeled from Chinese community question answering is chosen.

Paraphrase Identification (PI)

Paraphrase Identification aims to identify whether two sentences express the same meaning. We use cMedQQ, which consists of search query pairs.

Question Natural Language Inference (QNLI)

Question natural language inference aims to identify whether the answer corresponds to the question in the question-answering pair. We use cMedQNLI, which consists of question-answer pairs.

Question Answering (QA)

Question answering can be approximated as ranking candidate answer sentences based on their similarity. We assign 0,1 labels to the QA pairs, which convert to the binary classification problem. We use cMedQA released from the paper "Multi-Scale Attentive Interaction Networks for Chinese Medical Question Answer Selection", which consists of questions and their answers.

Information Retrieval (IR)

Information retrieval aims to retrieve most related documents given search queries. IR can be regarded as a ranking task. We use the cMedIR dataset, which consists of queries with multiple documents and their relative scores.

Intent Classification (IC)

Intent classification aims to assign intent labels to the queries, which can be regarded as multiple label classification tasks. We use the cMedIC dataset, which consists of queries with three intent labels (e.g., no intention, weak intention, and firm intention).

Text Classification (TC)

Text classification aims to assign multiple labels to the sentence. We use the cMedTC dataset, which consists of biomedical texts with multiple labels.

Symptom Diagnosis

Symptom diagnosis is a challenging yet profound problem in natural language processing. We use the CMDD dataset released from the paper"Enhancing Dialogue Symptom Diagnosis with Global Attention and Symptom Graph".

Relation extraction (RE)

Coming soon.

Datasets

All datasets can be downloaded at ChineseBLUE1.0

Pretrained Model

All pre-trained models can be downloaded at MC-BERT.

How to finetune?

sequence labeling task:

python finetune_cMedQANER.py\
  --task_name=cmedqaner\
  --do_train=True\
  --do_eval=True\
  --data_dir=./data/cMedQANER\
  --output_dir=model_cMedQANER\
  --bert_config_file=mc_bert_base/bert_config.json\
  --vocab_file=mc_bert_base/vocab.txt\
  --init_checkpoint=mc_bert_base/bert_model.ckpt

classification task:

python finetune_classifier.py\
  --task_name=cmedtc\
  --do_train=True\
  --do_eval=True\
  --data_dir=./data/cMedTC\
  --output_dir=model_cMedTC\
  --bert_config_file=mc_bert_base/bert_config.json\
  --vocab_file=mc_bert_base/vocab.txt\
  --init_checkpoint=mc_bert_base/bert_model.ckpt

Citing Chinese BLUE

Paper will be released soon.

@InProceedings{zhang2019cbert,
  author    = {Ningyu Zhang, Qianghuai Jia, Kangping Yin, Liang Dong, Feng Gao, Nengwei Hua},
  title     = {Conceptualized Representation Learning for Chinese Biomedical Text Mining},
  booktitle = {WSDM 2020},
  year      = {2020},
}

Acknowledgments

We are also grateful to the authors of BERT and wwm-BERT to make the data and codes publicly available. We are also grateful to the authors of paper "Enhancing Dialogue Symptom Diagnosis with Global Attention and Symptom Graph" and "Multi-Scale Attentive Interaction Networks for Chinese Medical Question Answer Selection" who released the dataset CMDD and cMedQA2.

Disclaimer

This project is not the official product of Alibaba. The information produced on this website is not intended for direct diagnostic use or medical decision-making without review and oversight by a clinical professional. Individuals should not change their health behavior solely on the basis of information produced on this website. If you have questions about the information produced on this website, please see a health care professional.

The experimental results presented in the technical report only indicate the performance under the specific data set and super-parameter combination and do not represent the essence of each model. The experimental results may change due to random number seeds and computing devices. The content of this project is for technical research purposes only and is not intended to be a conclusive basis. The user may use the model arbitrarily within the scope of the license, but we are not responsible for any direct or indirect damages resulting from the use of the content of the project.

About

Chinese Biomedical Language Understanding Evaluation benchmark (ChineseBLUE)

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published