Datasets:

ArXiv:
License:
mimicause / mimicause.py
imbesat-rizvi's picture
Emphasized on the Community Annotations Downloads section for dataset downloading
87615ea
import datasets
from pathlib import Path
import pandas as pd
from sklearn.model_selection import train_test_split
_DESCRIPTION = """\
MIMICause Dataset: A dataset for representation and automatic extraction of causal relation types from clinical notes.
The dataset has 2714 samples having both explicit and implicit causality in which entities are in the same sentence or different sentences.
The dataset has following nine semantic causal relations (with directionality) between entitities E1 and E2 in a text snippet:
(1) Cause(E1,E2)
(2) Cause(E2,E1)
(3) Enable(E1,E2)
(4) Enable(E2,E1)
(5) Prevent(E1,E2)
(6) Prevent(E2,E1)
(7) Hinder(E1,E2)
(8) Hinder(E2,E1)
(9) Other
"""
_DOWNLOAD_URL = "https://portal.dbmi.hms.harvard.edu/projects/n2c2-nlp/"
_CITATION = """\
@inproceedings{khetan-etal-2022-mimicause,
title={MIMICause: Representation and automatic extraction of causal relation types from clinical notes},
author={Vivek Khetan and Md Imbesat Hassan Rizvi and Jessica Huber and Paige Bartusiak and Bogdan Sacaleanu and Andrew Fano},
booktitle ={Findings of the Association for Computational Linguistics: ACL 2022},
month={may},
year={2022},
publisher={Association for Computational Linguistics},
address={Dublin, The Republic of Ireland},
url={},
doi={},
pages={},
}
"""
class MIMICAUSE(datasets.GeneratorBasedBuilder):
VERSION = datasets.utils.Version("1.0.0")
manual_download_instructions = (
"The MIMICause dataset requires manual download of the mimicause.zip "
"file from the Community Annotations Downloads of the DBMI Data Portal"
f" ({_DOWNLOAD_URL}) after signing their agreement forms, which is a "
"quick and easy procedure. Then provide the path of the downloaded "
"mimicause.zip file."
)
def _info(self):
features = datasets.Features(
{
"E1": datasets.Value("string"),
"E2": datasets.Value("string"),
"Text": datasets.Value("large_string"),
"Label": datasets.features.ClassLabel(
names=[
"Cause(E1,E2)",
"Cause(E2,E1)",
"Enable(E1,E2)",
"Enable(E2,E1)",
"Prevent(E1,E2)",
"Prevent(E2,E1)",
"Hinder(E1,E2)",
"Hinder(E2,E1)",
"Other",
],
),
}
)
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=features,
supervised_keys=None,
homepage=_DOWNLOAD_URL,
citation=_CITATION,
)
def _split_generators(self, dl_manager):
r"""
The dataset is split first in 90:10 ratio. The 90% split is further split
in 80:20 ratio. Thus the final split ratio is Train:Val:Test :: 72:18:10.
"""
filepath = dl_manager.download_and_extract(dl_manager.manual_dir)
filepath = Path(filepath) / "Annotations.csv"
data_df = pd.read_csv(filepath)
data_df = data_df.fillna("")
train_df, test_df = train_test_split(
data_df, test_size=0.1, stratify=data_df.Label, random_state=42
)
train_df, val_df = train_test_split(
train_df, test_size=0.2, stratify=train_df.Label, random_state=42
)
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN, gen_kwargs={"df": train_df}
),
datasets.SplitGenerator(
name=datasets.Split.VALIDATION, gen_kwargs={"df": val_df}
),
datasets.SplitGenerator(
name=datasets.Split.TEST, gen_kwargs={"df": test_df}
),
]
def _generate_examples(self, df):
for idx, row in df.iterrows():
yield idx, row.to_dict()