Mimic4Dataset / Mimic4Dataset.py
thbndi's picture
Update Mimic4Dataset.py
7cd7b81
raw
history blame
No virus
22.6 kB
import os
import pandas as pd
import datasets
import sys
import pickle
import subprocess
import shutil
from urllib.request import urlretrieve
from sklearn.model_selection import train_test_split
from sklearn.preprocessing import LabelEncoder
import yaml
from .dataset_utils import create_vocab,gender_vocab,vocab, concat_data, generate_deep, generate_ml
from .task_cohort import create_cohort, check_config_file
_DESCRIPTION = """\
Dataset for mimic4 data, by default for the Mortality task.
Available tasks are: Mortality, Length of Stay, Readmission, Phenotype.
The data is extracted from the mimic4 database using this pipeline: 'https://github.com/healthylaife/MIMIC-IV-Data-Pipeline/tree/main'
mimic path should have this form : "path/to/mimic4data/from/username/mimiciv/2.2"
If you choose a Custom task provide a configuration file for the Time series.
Currently working with Mimic-IV version 1 and 2
"""
_BASE_URL = "https://huggingface.co/datasets/thbndi/Mimic4Dataset/resolve/main"
_HOMEPAGE = "https://huggingface.co/datasets/thbndi/Mimic4Dataset"
_CITATION = "https://proceedings.mlr.press/v193/gupta22a.html"
_GIT_URL = "https://github.com/healthylaife/MIMIC-IV-Data-Pipeline"
_DATA_GEN = f"{_BASE_URL}/data_generation_icu_modify.py"
_DATA_GEN_HOSP= f"{_BASE_URL}/data_generation_modify.py"
_DAY_INT= f"{_BASE_URL}/day_intervals_cohort_v22.py"
_CONFIG_URLS = {'los' : f"{_BASE_URL}/config/los.config",
'mortality' : f"{_BASE_URL}/config/mortality.config",
'phenotype' : f"{_BASE_URL}/config/phenotype.config",
'readmission' : f"{_BASE_URL}/config/readmission.config"
}
class Mimic4DatasetConfig(datasets.BuilderConfig):
"""BuilderConfig for Mimic4Dataset."""
def __init__(
self,
**kwargs,
):
super().__init__(**kwargs)
class Mimic4Dataset(datasets.GeneratorBasedBuilder):
"""Create Mimic4Dataset dataset from Mimic-IV data stored in user machine."""
VERSION = datasets.Version("1.0.0")
def __init__(self, **kwargs):
self.mimic_path = kwargs.pop("mimic_path", None)
self.encoding = kwargs.pop("encoding",'concat')
self.config_path = kwargs.pop("config_path",None)
self.test_size = kwargs.pop("test_size",0.2)
self.val_size = kwargs.pop("val_size",0.1)
self.generate_cohort = kwargs.pop("generate_cohort",True)
if self.encoding == 'concat':
self.concat = True
else:
self.concat = False
super().__init__(**kwargs)
BUILDER_CONFIGS = [
Mimic4DatasetConfig(
name="Phenotype",
version=VERSION,
description="Dataset for mimic4 Phenotype task"
),
Mimic4DatasetConfig(
name="Readmission",
version=VERSION,
description="Dataset for mimic4 Readmission task"
),
Mimic4DatasetConfig(
name="Length of Stay",
version=VERSION,
description="Dataset for mimic4 Length of Stay task"
),
Mimic4DatasetConfig(
name="Mortality",
version=VERSION,
description="Dataset for mimic4 Mortality task"
),
]
DEFAULT_CONFIG_NAME = "Mortality"
def create_cohort(self):
if self.config_path==None:
if self.config.name == 'Phenotype' : self.config_path = _CONFIG_URLS['phenotype']
if self.config.name == 'Readmission' : self.config_path = _CONFIG_URLS['readmission']
if self.config.name == 'Length of Stay' : self.config_path = _CONFIG_URLS['los']
if self.config.name == 'Mortality' : self.config_path = _CONFIG_URLS['mortality']
version = self.mimic_path.split('/')[-1]
mimic_folder= self.mimic_path.split('/')[-2]
mimic_complete_path='/'+mimic_folder+'/'+version
current_directory = os.getcwd()
if os.path.exists(os.path.dirname(current_directory)+'/MIMIC-IV-Data-Pipeline-main'):
dir =os.path.dirname(current_directory)
os.chdir(dir)
else:
#move to parent directory of mimic data
dir = self.mimic_path.replace(mimic_complete_path,'')
if dir[-1]!='/':
dir=dir+'/'
elif dir=='':
dir="./"
parent_dir = os.path.dirname(self.mimic_path)
os.chdir(parent_dir)
#####################clone git repo if doesnt exists
repo_url='https://github.com/healthylaife/MIMIC-IV-Data-Pipeline'
if os.path.exists('MIMIC-IV-Data-Pipeline-main'):
path_bench = './MIMIC-IV-Data-Pipeline-main'
else:
path_bench ='./MIMIC-IV-Data-Pipeline-main'
subprocess.run(["git", "clone", repo_url, path_bench])
os.makedirs(path_bench+'/'+'mimic-iv')
shutil.move(version,path_bench+'/'+'mimic-iv')
os.chdir(path_bench)
self.mimic_path = './'+'mimic-iv'+'/'+version
####################Get configurations param
#download config file if not custom
if self.config_path[0:4] == 'http':
c = self.config_path.split('/')[-1]
file_path, head = urlretrieve(self.config_path,c)
else :
file_path = self.config_path
if not os.path.exists('./config'):
os.makedirs('config')
#save config file in config folder
self.conf='./config/'+file_path.split('/')[-1]
if not os.path.exists(self.conf):
shutil.move(file_path,'./config')
with open(self.conf) as f:
config = yaml.safe_load(f)
timeW = config['timeWindow']
self.timeW=int(timeW.split()[1])
self.bucket = config['timebucket']
self.data_icu = config['icu_no_icu']=='ICU'
if self.data_icu:
self.feat_cond, self.feat_chart, self.feat_proc, self.feat_meds, self.feat_out, self.lab = config['diagnosis'], config['chart'], config['proc'], config['meds'], config['output'], False
self.feat_lab = False
else:
self.feat_cond, self.feat_lab, self.feat_proc, self.feat_meds, self.feat_chart, self.out = config['diagnosis'], config['lab'], config['proc'], config['meds'], False, False
self.feat_out = False
self.feat_chart = False
#####################downloads modules from hub
if not os.path.exists('./model/data_generation_icu_modify.py'):
file_path, head = urlretrieve(_DATA_GEN, "data_generation_icu_modify.py")
shutil.move(file_path, './model')
if not os.path.exists('./model/data_generation_modify.py'):
file_path, head = urlretrieve(_DATA_GEN_HOSP, "data_generation_modify.py")
shutil.move(file_path, './model')
if not os.path.exists('./preprocessing/day_intervals_preproc/day_intervals_cohort_v22.py'):
file_path, head = urlretrieve(_DAY_INT, "day_intervals_cohort_v22.py")
shutil.move(file_path, './preprocessing/day_intervals_preproc')
data_dir = "./data/dict/"+self.config.name.replace(" ","_")+"/dataDic"
sys.path.append(path_bench)
config = self.config_path.split('/')[-1]
#####################create task cohort
if self.generate_cohort:
create_cohort(self.config.name.replace(" ","_"),self.mimic_path,config)
#####################Split data into train, test and val
with open(data_dir, 'rb') as fp:
dataDic = pickle.load(fp)
data = pd.DataFrame.from_dict(dataDic)
dict_dir = "./data/dict/"+self.config.name.replace(" ","_")
data=data.T
train_data, test_data = train_test_split(data, test_size=self.test_size, random_state=42)
if self.val_size > 0 :
train_data, val_data = train_test_split(train_data, test_size=self.val_size, random_state=42)
val_dic = val_data.to_dict('index')
val_path = dict_dir+'/val_data.pkl'
with open(val_path, 'wb') as f:
pickle.dump(val_dic, f)
train_dic = train_data.to_dict('index')
test_dic = test_data.to_dict('index')
train_path = dict_dir+'/train_data.pkl'
test_path = dict_dir+'/test_data.pkl'
with open(train_path, 'wb') as f:
pickle.dump(train_dic, f)
with open(test_path, 'wb') as f:
pickle.dump(test_dic, f)
return dict_dir
###########################################################RAW##################################################################
def _info_raw(self):
features = datasets.Features(
{
"label": datasets.ClassLabel(num_classes=2,names=["0", "1"]),
"gender": datasets.Value("string"),
"ethnicity": datasets.Value("string"),
"insurance": datasets.Value("string"),
"age": datasets.Value("int32"),
"COND": datasets.Sequence(datasets.Value("string")),
"MEDS": {
"signal":
{
"id": datasets.Sequence(datasets.Value("int32")),
"value": datasets.Sequence(datasets.Sequence(datasets.Value("float32")))
}
,
"rate":
{
"id": datasets.Sequence(datasets.Value("int32")),
"value": datasets.Sequence(datasets.Sequence(datasets.Value("float32")))
}
,
"amount":
{
"id": datasets.Sequence(datasets.Value("int32")),
"value": datasets.Sequence(datasets.Sequence(datasets.Value("float32")))
}
},
"PROC": {
"id": datasets.Sequence(datasets.Value("int32")),
"value": datasets.Sequence(datasets.Sequence(datasets.Value("float32")))
},
"CHART/LAB":
{
"signal" : {
"id": datasets.Sequence(datasets.Value("int32")),
"value": datasets.Sequence(datasets.Sequence(datasets.Value("float32")))
},
"val" : {
"id": datasets.Sequence(datasets.Value("int32")),
"value": datasets.Sequence(datasets.Sequence(datasets.Value("float32")))
},
},
"OUT": {
"id": datasets.Sequence(datasets.Value("int32")),
"value": datasets.Sequence(datasets.Sequence(datasets.Value("float32")))
},
}
)
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=features,
homepage=_HOMEPAGE,
citation=_CITATION,
)
def _generate_examples_raw(self, filepath):
with open(filepath, 'rb') as fp:
dataDic = pickle.load(fp)
for hid, data in dataDic.items():
proc_features = data['Proc']
meds_features = data['Med']
out_features = data['Out']
cond_features = data['Cond']['fids']
eth= data['ethnicity']
age = data['age']
gender = data['gender']
label = data['label']
insurance=data['insurance']
items = list(proc_features.keys())
values =[proc_features[i] for i in items ]
procs = {"id" : items,
"value": values}
items_outs = list(out_features.keys())
values_outs =[out_features[i] for i in items_outs ]
outs = {"id" : items_outs,
"value": values_outs}
if self.data_icu:
chart_features = data['Chart']
else:
chart_features = data['Lab']
#chart signal
if ('signal' in chart_features):
items_chart_sig = list(chart_features['signal'].keys())
values_chart_sig =[chart_features['signal'][i] for i in items_chart_sig ]
chart_sig = {"id" : items_chart_sig,
"value": values_chart_sig}
else:
chart_sig = {"id" : [],
"value": []}
#chart val
if ('val' in chart_features):
items_chart_val = list(chart_features['val'].keys())
values_chart_val =[chart_features['val'][i] for i in items_chart_val ]
chart_val = {"id" : items_chart_val,
"value": values_chart_val}
else:
chart_val = {"id" : [],
"value": []}
charts = {"signal" : chart_sig,
"val" : chart_val}
#meds signal
if ('signal' in meds_features):
items_meds_sig = list(meds_features['signal'].keys())
values_meds_sig =[meds_features['signal'][i] for i in items_meds_sig ]
meds_sig = {"id" : items_meds_sig,
"value": values_meds_sig}
else:
meds_sig = {"id" : [],
"value": []}
#meds rate
if ('rate' in meds_features):
items_meds_rate = list(meds_features['rate'].keys())
values_meds_rate =[meds_features['rate'][i] for i in items_meds_rate ]
meds_rate = {"id" : items_meds_rate,
"value": values_meds_rate}
else:
meds_rate = {"id" : [],
"value": []}
#meds amount
if ('amount' in meds_features):
items_meds_amount = list(meds_features['amount'].keys())
values_meds_amount =[meds_features['amount'][i] for i in items_meds_amount ]
meds_amount = {"id" : items_meds_amount,
"value": values_meds_amount}
else:
meds_amount = {"id" : [],
"value": []}
meds = {"signal" : meds_sig,
"rate" : meds_rate,
"amount" : meds_amount}
yield int(hid), {
"label" : label,
"gender" : gender,
"ethnicity" : eth,
"insurance" : insurance,
"age" : age,
"COND" : cond_features,
"PROC" : procs,
"CHART/LAB" : charts,
"OUT" : outs,
"MEDS" : meds
}
###########################################################ENCODED##################################################################
def _info_encoded(self):
features = datasets.Features(
{
"label": datasets.ClassLabel(num_classes=2,names=["0", "1"]),
"features" : datasets.Sequence(datasets.Value("float32")),
}
)
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=features,
homepage=_HOMEPAGE,
citation=_CITATION,
)
def _generate_examples_encoded(self, filepath):
path= './data/dict/'+self.config.name.replace(" ","_")+'/ethVocab'
with open(path, 'rb') as fp:
ethVocab = pickle.load(fp)
path= './data/dict/'+self.config.name.replace(" ","_")+'/insVocab'
with open(path, 'rb') as fp:
insVocab = pickle.load(fp)
genVocab = ['<PAD>', 'M', 'F']
gen_encoder = LabelEncoder()
eth_encoder = LabelEncoder()
ins_encoder = LabelEncoder()
gen_encoder.fit(genVocab)
eth_encoder.fit(ethVocab)
ins_encoder.fit(insVocab)
with open(filepath, 'rb') as fp:
dico = pickle.load(fp)
df = pd.DataFrame.from_dict(dico, orient='index')
task=self.config.name.replace(" ","_")
for i, data in df.iterrows():
concat_cols=[]
dyn_df,cond_df,demo=concat_data(data,task,self.feat_cond,self.feat_proc,self.feat_out, self.feat_chart, self.feat_meds,self.feat_lab)
dyn=dyn_df.copy()
dyn.columns=dyn.columns.droplevel(0)
cols=dyn.columns
time=dyn.shape[0]
for t in range(time):
cols_t = [str(x) + "_"+str(t) for x in cols]
concat_cols.extend(cols_t)
demo['gender']=gen_encoder.transform(demo['gender'])
demo['ethnicity']=eth_encoder.transform(demo['ethnicity'])
demo['insurance']=ins_encoder.transform(demo['insurance'])
label = data['label']
demo=demo.drop(['label'],axis=1)
X= generate_ml(dyn_df,cond_df,demo,concat_cols,self.concat)
X=X.values.tolist()[0]
yield int(i), {
"label": label,
"features": X,
}
######################################################DEEP###############################################################
def _info_deep(self):
features = datasets.Features(
{
"label": datasets.ClassLabel(num_classes=2,names=["0", "1"]),
"DEMO": datasets.Sequence(datasets.Value("int64")),
"COND" : datasets.Sequence(datasets.Value("int64")),
"MEDS" : datasets.Array2D(shape=(None, self.size_meds), dtype='int64') ,
"PROC" : datasets.Array2D(shape=(None, self.size_proc), dtype='int64') ,
"CHART/LAB" : datasets.Array2D(shape=(None, self.size_chart), dtype='int64') ,
"OUT" : datasets.Array2D(shape=(None, self.size_out), dtype='int64') ,
}
)
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=features,
homepage=_HOMEPAGE,
citation=_CITATION,
)
def _generate_examples_deep(self, filepath):
with open(filepath, 'rb') as fp:
dico = pickle.load(fp)
task=self.config.name.replace(" ","_")
for key, data in dico.items():
stat, demo, meds, chart, out, proc, lab, y = generate_deep(data, task, self.feat_cond, self.feat_proc, self.feat_out, self.feat_chart, self.feat_meds,self.feat_lab)
verri=True
if self.feat_proc:
if (len(proc)<(self.timeW//self.bucket)):
verri=False
if self.feat_out:
if (len(out)<(self.timeW//self.bucket)):
verri=False
if self.feat_chart:
if (len(chart)<(self.timeW//self.bucket)):
verri=False
if self.feat_meds:
if (len(meds)<(self.timeW//self.bucket)):
verri=False
if self.feat_lab:
if (len(lab)<(self.timeW//self.bucket)):
verri=False
if verri:
if self.data_icu:
yield int(key), {
'label': y,
'DEMO': demo,
'COND': stat,
'MEDS': meds,
'PROC': proc,
'CHART/LAB': chart,
'OUT': out,
}
else:
yield int(key), {
'label': y,
'DEMO': demo,
'COND': stat,
'MEDS': meds,
'PROC': proc,
'CHART/LAB': lab,
'OUT': out,
}
else:
continue
#############################################################################################################################
def _info(self):
self.path = self.create_cohort()
self.size_cond, self.size_proc, self.size_meds, self.size_out, self.size_chart, self.size_lab, eth_vocab,gender_vocab,age_vocab,ins_vocab=vocab(self.config.name.replace(" ","_"),self.feat_cond,self.feat_proc,self.feat_out,self.feat_chart,self.feat_meds,self.feat_lab)
if self.encoding == 'concat' :
return self._info_encoded()
elif self.encoding == 'aggreg' :
return self._info_encoded()
elif self.encoding == 'tensor' :
return self._info_deep()
else:
return self._info_raw()
def _split_generators(self, dl_manager):
csv_dir = "./data/dict/"+self.config.name.replace(" ","_")
if self.val_size > 0 :
return [
datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": csv_dir+'/train_data.pkl'}),
datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={"filepath": csv_dir+'/val_data.pkl'}),
datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"filepath": csv_dir+'/test_data.pkl'}),
]
else :
return [
datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": csv_dir+'/train_data.pkl'}),
datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"filepath": csv_dir+'/test_data.pkl'}),
]
def _generate_examples(self, filepath):
if self.encoding == 'concat' :
yield from self._generate_examples_encoded(filepath)
elif self.encoding == 'aggreg' :
yield from self._generate_examples_encoded(filepath)
elif self.encoding == 'tensor' :
yield from self._generate_examples_deep(filepath)
else :
yield from self._generate_examples_raw(filepath)