All Projects → ymcui → expmrc

ymcui / expmrc

Licence: CC-BY-SA-4.0 license
ExpMRC: Explainability Evaluation for Machine Reading Comprehension

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to expmrc

mllp
The code of AAAI 2020 paper "Transparent Classification with Multilayer Logical Perceptrons and Random Binarization".
Stars: ✭ 15 (-74.14%)
Mutual labels:  explainable-ai, xai
awesome-graph-explainability-papers
Papers about explainability of GNNs
Stars: ✭ 153 (+163.79%)
Mutual labels:  explainable-ai, xai
dlime experiments
In this work, we propose a deterministic version of Local Interpretable Model Agnostic Explanations (LIME) and the experimental results on three different medical datasets shows the superiority for Deterministic Local Interpretable Model-Agnostic Explanations (DLIME).
Stars: ✭ 21 (-63.79%)
Mutual labels:  explainable-ai, xai
concept-based-xai
Library implementing state-of-the-art Concept-based and Disentanglement Learning methods for Explainable AI
Stars: ✭ 41 (-29.31%)
Mutual labels:  explainable-ai, xai
verseagility
Ramp up your custom natural language processing (NLP) task, allowing you to bring your own data, use your preferred frameworks and bring models into production.
Stars: ✭ 23 (-60.34%)
Mutual labels:  question-answering, machine-reading-comprehension
fastshap
Fast approximate Shapley values in R
Stars: ✭ 79 (+36.21%)
Mutual labels:  explainable-ai, xai
zennit
Zennit is a high-level framework in Python using PyTorch for explaining/exploring neural networks using attribution methods like LRP.
Stars: ✭ 57 (-1.72%)
Mutual labels:  explainable-ai, xai
unanswerable qa
The official implementation for ACL 2021 "Challenges in Information Seeking QA: Unanswerable Questions and Paragraph Retrieval".
Stars: ✭ 21 (-63.79%)
Mutual labels:  question-answering, machine-reading-comprehension
Interpret
Fit interpretable models. Explain blackbox machine learning.
Stars: ✭ 4,352 (+7403.45%)
Mutual labels:  explainable-ai, xai
ml-fairness-framework
FairPut - Machine Learning Fairness Framework with LightGBM — Explainability, Robustness, Fairness (by @firmai)
Stars: ✭ 59 (+1.72%)
Mutual labels:  explainable-ai, xai
strategy
Improving Machine Reading Comprehension with General Reading Strategies
Stars: ✭ 35 (-39.66%)
Mutual labels:  question-answering, machine-reading-comprehension
mindsdb server
MindsDB server allows you to consume and expose MindsDB workflows, through http.
Stars: ✭ 3 (-94.83%)
Mutual labels:  explainable-ai, xai
mindsdb native
Machine Learning in one line of code
Stars: ✭ 34 (-41.38%)
Mutual labels:  xai
examinee
Laravel Quiz and Exam System clone of udemy
Stars: ✭ 151 (+160.34%)
Mutual labels:  question-answering
FinBERT-QA
Financial Domain Question Answering with pre-trained BERT Language Model
Stars: ✭ 70 (+20.69%)
Mutual labels:  question-answering
VideoNavQA
An alternative EQA paradigm and informative benchmark + models (BMVC 2019, ViGIL 2019 spotlight)
Stars: ✭ 22 (-62.07%)
Mutual labels:  question-answering
fastT5
⚡ boost inference speed of T5 models by 5x & reduce the model size by 3x.
Stars: ✭ 421 (+625.86%)
Mutual labels:  question-answering
cmrc2019
A Sentence Cloze Dataset for Chinese Machine Reading Comprehension (CMRC 2019)
Stars: ✭ 118 (+103.45%)
Mutual labels:  question-answering
DrFAQ
DrFAQ is a plug-and-play question answering NLP chatbot that can be generally applied to any organisation's text corpora.
Stars: ✭ 29 (-50%)
Mutual labels:  question-answering
datafsm
Machine Learning Finite State Machine Models from Data with Genetic Algorithms
Stars: ✭ 14 (-75.86%)
Mutual labels:  explainable-ai



GitHub

With the development of the pre-trained language models (PLMs), achieving human-level performance on several machine reading comprehension (MRC) dataset is not as hard as it used to be. However, the explainability behind these artifacts still remains unclear, raising concerns on utilizing these models in real-life applications. To improve the explainability of MRC tasks, we propose ExpMRC benchmark.

ExpMRC is a benchmark for Explainability Evaluation of Machine Reading Comprehension. ExpMRC contains four subsets of popular MRC datasets with additionally annotated evidences, including SQuAD, CMRC 2018, RACE+ (similar to RACE), and C3, covering span-extraction and multiple-choice questions MRC tasks in both English and Chinese.

To achieve a higher score in ExpMRC, the model should not only give a correct answer for the question but also give a passage span as the evidence text. We greatly welcome the submission that could be generalized well on different languages and types of MRC tasks with unsupervised or semi-supervised approaches.

ExpMRC: Explainability Evaluation for Machine Reading Comprehension

[Official Publication] [arXiv pre-print] [Leaderboard] [Papers With Code]

News

[Apri 19, 2022] Our paper is officially published in Heliyon.

[June 22, 2021] Baseline codes and pseudo training data are available, check baseline and pseudo-training-data directory.

[May 24, 2021] We have released our dataset, check data directory. The submission site is also open.

[May 17, 2021] Thank you for your interest in our dataset. We are about to release the dataset and baseline codes in the next few weeks, hopefully on late May. Stay tuned!

Directory Guide

| -- **ExpMRC Root**  
    | -- baseline               # Baseline codes  
    | -- data                   # ExpMRC development sets  
    | -- pseudo_training_data   # Pseudo training data  
    | -- sample_submission      # Sample submission files for ExpMRC  
    | -- eval_expmrc.py         # Evaluation script  

As stated in the paper, we DO NOT provide any training data. We intend to encourage our community to develop unsupervised or semi-supervised approaches for promoting Explainable MRC. Nonetheless, we provide the pseudo training data that was used in our paper. Please check pseudo-training-data directory.

Submission to Leaderboard

Please visit our leaderboard for more information: https://ymcui.github.io/expmrc/

To preserve the integrity of test results and improve the reproducibility, we do not release the test sets to the public. Instead, we require you to upload your model onto CodaLab, so that we can run it on the test sets for you. You can follow the instructions on CodaLab (which is similar to SQuAD, CMRC 2018 submission). You can submit your model on one or more subsets in ExpMRC. Sample submission files are shown in sample_submission directory.

Submission policies:

  1. You are free to use any open-source MRC data or automatically generated data for training your systems (both labeled and unlabeled).
  2. You are NOT allowed to use any publicly unavailable human-annotated data for training.
  3. We do not encourage using the development set of ExpMRC for training (though it is not prohibited). You should declare whether the system is trained by using the whole/part of the development set. Such submissions will be marked with an asterisk (*).

Citation

If you are using our benchmark in your work, please cite:

@article{cui-etal-2022-expmrc,
  title={ExpMRC: Explainability Evaluation for Machine Reading Comprehension},
  author={Cui, Yiming and Liu, Ting and Che, Wanxiang and Chen, Zhigang and Wang, Shijin},
  journal={Heliyon},
  year={2022},
  volume={8},
  issue={4},
  pages={e09290},
  issn={2405-8440},
  doi={https://doi.org/10.1016/j.heliyon.2022.e09290}
}

Acknowledgment

Yiming Cui would like to thank Google TPU Research Cloud (TRC) program for providing computing resource. We also thank SQuAD team for open-sourcing their website template.

Contact us

Please submit an issue.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].