File size: 5,762 Bytes
45a0093
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
from glob import glob
from pathlib import Path

import json
import numpy as np
import pandas as pd

"""
Dataset url: https://zenodo.org/record/3936490/files/annotated_corpus.zip?download=1
Paper url: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9044329/

There are no splits available ==> Make random split ourselves

"""

pd.set_option('display.max_colwidth', None)
pd.set_option('display.max_columns', None)


def analyze_dataset(df, num_characters_for_short_sentence=25):
    short_sentence = False
    counter = 0
    same_label_counter = 0
    other_number = 0
    num_one_paragraph_len = 0
    for i in range(200):
        for paragraph in df.iloc[i].decision_reasons:
            for sentence in paragraph:
                if short_sentence:
                    print("previous sentence was short: ", short_sentence)
                    print("current sentence label: ", sentence[1])
                    print("current paragraph: ", paragraph)
                    if sentence[1] == short_sentence[1]:
                        same_label_counter += 1

                if len(sentence[0]) < num_characters_for_short_sentence:
                    counter += 1
                    short_sentence = sentence
                    print()
                    print("short sentence: ", sentence)
                    print("short paragraph: ", paragraph)
                    if sentence[1] == 'other':
                        other_number += 1
                    if len(paragraph) == 1:
                        num_one_paragraph_len += 1
                else:
                    short_sentence = False

    print("num short sentences: ", counter)
    print("num short sentences containing the same label as the next one: ", same_label_counter)
    print("num short sentences containing 'other' as label: ", other_number)
    print("num short sentences where the paragraph contains only this one short sentence: ", num_one_paragraph_len)
    # ==> the label is only the same in very few cases
    # ==> the label is 'other' in the majority of cases, when it is not: it seems to be mislabeled
    # ==> think about removing them entirely
    # ==> we opted for not interfering in the content of the dataset


# create a summary jsonl file
dataset_filename = "dataset.jsonl"
if not Path(dataset_filename).exists():
    with open(dataset_filename, "a") as dataset_file:
        for filename in glob("annotated_corpus/*.json"):
            # we need to do this charade, because some jsons are formatted differently than others
            json_text = Path(filename).read_text()
            json_obj = json.loads(json_text)
            # make it less nested so that it is easier to read as df
            new_dict = {}
            new_dict.update(json_obj["meta"])
            new_dict.update(json_obj["decision_text"])
            dataset_file.write(json.dumps(new_dict) + "\n")
else:
    print(f"{dataset_filename} already exists. Please delete it to re-aggregate it.")

df = pd.read_json(dataset_filename, lines=True)

# Do splits before expanding the df so that entire decisions are in the splits and not samples from one decision are spread across splits
# perform random split 80% train (160 decisions), 10% validation (20 decisions), 10% test (20 decisions)
train, validation, test = np.split(df.sample(frac=1, random_state=42), [int(.8 * len(df)), int(.9 * len(df))])


def expand_df(df):
    """
    Expand the df so that each sentence has its own row and is its own sample
    :param df:
    :return:
    """
    rows = []
    for index, row in df.iterrows():
        for paragraph in row.decision_reasons:
            for sent_idx, sentence in enumerate(paragraph):
                new_row = {'file_number': row['file_number'], 'input_sentence': sentence[0], 'label': sentence[1]}
                # Discussion with lawyer yielded, that the paragraph as context is enough
                # take the sentences before
                new_row['context_before'] = paragraph[:sent_idx]
                # take the remaining sentences afterwards
                new_row['context_after'] = paragraph[sent_idx + 1:]
                rows.append(new_row)

    return pd.DataFrame.from_records(rows)


train = expand_df(train)
validation = expand_df(validation)
test = expand_df(test)

# Num samples for each split: train (19271), validation (2726), test (3078)
print(len(train.index), len(validation.index), len(test.index))

# save to jsonl files for huggingface
train.to_json("train.jsonl", lines=True, orient="records")
validation.to_json("validation.jsonl", lines=True, orient="records")
test.to_json("test.jsonl", lines=True, orient="records")

# save main df with meta information to file
# link to splits is given via file_number
df = df.drop(['decision_reasons'], axis=1)
df.to_json("meta.jsonl", lines=True, orient="records")


def print_split_table_single_label(train, validation, test, label_name):
    train_counts = train[label_name].value_counts().to_frame().rename(columns={label_name: "train"})
    validation_counts = validation[label_name].value_counts().to_frame().rename(columns={label_name: "validation"})
    test_counts = test[label_name].value_counts().to_frame().rename(columns={label_name: "test"})

    table = train_counts.join(validation_counts)
    table = table.join(test_counts)
    table[label_name] = table.index
    total_row = {label_name: "total",
                 "train": len(train.index),
                 "validation": len(validation.index),
                 "test": len(test.index)}
    table = table.append(total_row, ignore_index=True)
    table = table[[label_name, "train", "validation", "test"]]  # reorder columns
    print(table.to_markdown(index=False))


print_split_table_single_label(train, validation, test, "label")