Joelito's picture
simplified the dataset to avoid errors
176a3f1
raw history blame
No virus
1.58 kB
import pandas as pd
import os
from typing import Union
import datasets
from datasets import load_dataset
def save_and_compress(dataset: Union[datasets.Dataset, pd.DataFrame], name: str, idx=None):
if idx:
path = f"{name}_{idx}.jsonl"
else:
path = f"{name}.jsonl"
print("Saving to", path)
dataset.to_json(path, force_ascii=False, orient='records', lines=True)
print("Compressing...")
os.system(f'xz -zkf -T0 {path}') # -TO to use multithreading
def get_dataset_column_from_text_folder(folder_path):
return load_dataset("text", data_dir=folder_path, sample_by="document", split='train').to_pandas()['text']
for split in ["train", "test"]:
dfs = []
for dataset_name in ["IN-Abs", "UK-Abs", "IN-Ext"]:
if dataset_name == "IN-Ext" and split == "test":
continue
print(f"Processing {dataset_name} {split}")
path = f"original_dataset/{dataset_name}/{split}-data"
df = pd.DataFrame()
df['judgement'] = get_dataset_column_from_text_folder(f"{path}/judgement")
df['dataset_name'] = dataset_name
if dataset_name == "UK-Abs" and split == "test" or dataset_name == "IN-Ext":
summary_full_path = f"{path}/summary/full"
else:
summary_full_path = f"{path}/summary"
df['summary'] = get_dataset_column_from_text_folder(summary_full_path)
dfs.append(df)
df = pd.concat(dfs)
df = df.fillna("") # NaNs can lead to huggingface not recognizing the feature type of the column
save_and_compress(df, f"data/{split}")