Datasets:
Tasks:
Text Classification
Modalities:
Text
Formats:
json
Sub-tasks:
multi-class-classification
Languages:
German
Size:
10K<n<100K
Tags:
License:
File size: 5,762 Bytes
45a0093 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 |
from glob import glob
from pathlib import Path
import json
import numpy as np
import pandas as pd
"""
Dataset url: https://zenodo.org/record/3936490/files/annotated_corpus.zip?download=1
Paper url: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9044329/
There are no splits available ==> Make random split ourselves
"""
pd.set_option('display.max_colwidth', None)
pd.set_option('display.max_columns', None)
def analyze_dataset(df, num_characters_for_short_sentence=25):
short_sentence = False
counter = 0
same_label_counter = 0
other_number = 0
num_one_paragraph_len = 0
for i in range(200):
for paragraph in df.iloc[i].decision_reasons:
for sentence in paragraph:
if short_sentence:
print("previous sentence was short: ", short_sentence)
print("current sentence label: ", sentence[1])
print("current paragraph: ", paragraph)
if sentence[1] == short_sentence[1]:
same_label_counter += 1
if len(sentence[0]) < num_characters_for_short_sentence:
counter += 1
short_sentence = sentence
print()
print("short sentence: ", sentence)
print("short paragraph: ", paragraph)
if sentence[1] == 'other':
other_number += 1
if len(paragraph) == 1:
num_one_paragraph_len += 1
else:
short_sentence = False
print("num short sentences: ", counter)
print("num short sentences containing the same label as the next one: ", same_label_counter)
print("num short sentences containing 'other' as label: ", other_number)
print("num short sentences where the paragraph contains only this one short sentence: ", num_one_paragraph_len)
# ==> the label is only the same in very few cases
# ==> the label is 'other' in the majority of cases, when it is not: it seems to be mislabeled
# ==> think about removing them entirely
# ==> we opted for not interfering in the content of the dataset
# create a summary jsonl file
dataset_filename = "dataset.jsonl"
if not Path(dataset_filename).exists():
with open(dataset_filename, "a") as dataset_file:
for filename in glob("annotated_corpus/*.json"):
# we need to do this charade, because some jsons are formatted differently than others
json_text = Path(filename).read_text()
json_obj = json.loads(json_text)
# make it less nested so that it is easier to read as df
new_dict = {}
new_dict.update(json_obj["meta"])
new_dict.update(json_obj["decision_text"])
dataset_file.write(json.dumps(new_dict) + "\n")
else:
print(f"{dataset_filename} already exists. Please delete it to re-aggregate it.")
df = pd.read_json(dataset_filename, lines=True)
# Do splits before expanding the df so that entire decisions are in the splits and not samples from one decision are spread across splits
# perform random split 80% train (160 decisions), 10% validation (20 decisions), 10% test (20 decisions)
train, validation, test = np.split(df.sample(frac=1, random_state=42), [int(.8 * len(df)), int(.9 * len(df))])
def expand_df(df):
"""
Expand the df so that each sentence has its own row and is its own sample
:param df:
:return:
"""
rows = []
for index, row in df.iterrows():
for paragraph in row.decision_reasons:
for sent_idx, sentence in enumerate(paragraph):
new_row = {'file_number': row['file_number'], 'input_sentence': sentence[0], 'label': sentence[1]}
# Discussion with lawyer yielded, that the paragraph as context is enough
# take the sentences before
new_row['context_before'] = paragraph[:sent_idx]
# take the remaining sentences afterwards
new_row['context_after'] = paragraph[sent_idx + 1:]
rows.append(new_row)
return pd.DataFrame.from_records(rows)
train = expand_df(train)
validation = expand_df(validation)
test = expand_df(test)
# Num samples for each split: train (19271), validation (2726), test (3078)
print(len(train.index), len(validation.index), len(test.index))
# save to jsonl files for huggingface
train.to_json("train.jsonl", lines=True, orient="records")
validation.to_json("validation.jsonl", lines=True, orient="records")
test.to_json("test.jsonl", lines=True, orient="records")
# save main df with meta information to file
# link to splits is given via file_number
df = df.drop(['decision_reasons'], axis=1)
df.to_json("meta.jsonl", lines=True, orient="records")
def print_split_table_single_label(train, validation, test, label_name):
train_counts = train[label_name].value_counts().to_frame().rename(columns={label_name: "train"})
validation_counts = validation[label_name].value_counts().to_frame().rename(columns={label_name: "validation"})
test_counts = test[label_name].value_counts().to_frame().rename(columns={label_name: "test"})
table = train_counts.join(validation_counts)
table = table.join(test_counts)
table[label_name] = table.index
total_row = {label_name: "total",
"train": len(train.index),
"validation": len(validation.index),
"test": len(test.index)}
table = table.append(total_row, ignore_index=True)
table = table[[label_name, "train", "validation", "test"]] # reorder columns
print(table.to_markdown(index=False))
print_split_table_single_label(train, validation, test, "label")
|