MedExpQA / README.md
ragerri's picture
Update README.md
f6bc49f verified
metadata
license: cc-by-4.0
language:
  - en
  - es
  - fr
  - it
tags:
  - casimedicos
  - explainability
  - medical exams
  - medical question answering
  - multilinguality
  - LLMs
  - LLM
pretty_name: MedExpQA
configs:
  - config_name: en
    data_files:
      - split: train
        path:
          - data/en/train.en.casimedicos.rag.jsonl
      - split: validation
        path:
          - data/en/dev.en.casimedicos.rag.jsonl
      - split: test
        path:
          - data/en/test.en.casimedicos.rag.jsonl
  - config_name: es
    data_files:
      - split: train
        path:
          - data/es/train.es.casimedicos.rag.jsonl
      - split: validation
        path:
          - data/es/dev.es.casimedicos.rag.jsonl
      - split: test
        path:
          - data/es/test.es.casimedicos.rag.jsonl
  - config_name: fr
    data_files:
      - split: train
        path:
          - data/fr/train.fr.casimedicos.rag.jsonl
      - split: validation
        path:
          - data/fr/dev.fr.casimedicos.rag.jsonl
      - split: test
        path:
          - data/fr/test.fr.casimedicos.rag.jsonl
  - config_name: it
    data_files:
      - split: train
        path:
          - data/it/train.it.casimedicos.rag.jsonl
      - split: validation
        path:
          - data/it/dev.it.casimedicos.rag.jsonl
      - split: test
        path:
          - data/it/test.it.casimedicos.rag.jsonl
task_categories:
  - text-generation
  - question-answering
size_categories:
  - 1K<n<10K



MexExpQA: Multilingual Benchmarking of Medical QA with reference gold explanations and Retrieval Augmented Generation (RAG)

We present a new multilingual parallel medical benchmark, MedExpQA, for the evaluation of LLMs on Medical Question Answering. This benchmark can be used for various NLP tasks including: Medical Question Answering or Explanation Generation.

Although the design of MedExpQA is independent of any specific dataset, for the first version of the MedExpQA benchmark we leverage the commented MIR exams from the Antidote CasiMedicos dataset which includes gold reference explanations, which is currently available for 4 languages: English, French, Italian and Spanish.

Antidote CasiMedicos splits
train 434
validation 63
test 125

Example of Document in Antidote CasiMedicos Dataset

In this repository you can find the following data:

  • casimedicos-raw: The textual content including Clinical Case (C), Question (Q), Possible Answers (P), and Explanation (E) as shown in the example above.
  • casimedicos-exp: The manual annotations linking the explanations of the correct and incorrect possible answers.
  • MedExpQA: benchmark for Medical QA based on gold reference explanations from casimedicos-exp and knowledge automatically extracted using RAG methods.

Data Explanation

The following attributes composed casimedicos-raw:

  • id: unique doc identifier.
  • year: year in which the exam was published by the Spanish Ministry of Health.
  • question_id_specific: id given to the original exam published by the Spanish Ministry of Health.
  • full_question: Clinical Case (C) and Question (Q) as illustrated in the example document above.
  • full answer: Full commented explanation (E) as illustrated in the example document above.
  • type: medical speciality.
  • options: Possible Answers (P) as illustrated in the example document above.
  • correct option: solution to the exam question.

Additionally, the following jsonl attribute was added to create casimedicos-exp:

  • explanations: for each possible answer above, manual annotation states whether:
    1. the explanation for each possible answer exists in the full comment (E) and
    2. if present, then we provide character and token offsets plus the text corresponding to the explanation for each possible answer.

For MedExpQA benchmarking we have added the following elements in the data:

  • rag
    1. clinical_case_options/MedCorp/RRF-2: 32 snippets extracted from the MedCorp corpus using the combination of clinical case and options as a query during the retrieval process. These 32 snippets are the resulting RRF combination of 32 separately retrieved snippets using BM25 and MedCPT.

MedExpQA Benchmark Overview

Prompt Example for LLMs

Benchmark Results (averaged per type of external knowledge for grounding)

LLMs evaluated: LLaMA, PMC-LLaMA, Mistral and BioMistral.

Citation

If you use MedExpQA data then please cite the following papers:

@misc{alonso2024medexpqa,
      title={MedExpQA: Multilingual Benchmarking of Large Language Models for Medical Question Answering}, 
      author={Iñigo Alonso and Maite Oronoz and Rodrigo Agerri},
      year={2024},
      eprint={2404.05590},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

@misc{goenaga2023explanatory,
      title={Explanatory Argument Extraction of Correct Answers in Resident Medical Exams}, 
      author={Iakes Goenaga and Aitziber Atutxa and Koldo Gojenola and Maite Oronoz and Rodrigo Agerri},
      year={2023},
      eprint={2312.00567},
      archivePrefix={arXiv}
}

Contact: Iñigo Alonso and Rodrigo Agerri HiTZ Center - Ixa, University of the Basque Country UPV/EHU