File size: 1,538 Bytes
36baa0d
 
 
 
 
7ad0deb
 
 
 
 
 
 
 
 
 
 
 
6b445fb
 
 
7ad0deb
f681059
 
 
7ad0deb
 
 
f681059
 
c518314
 
f681059
6b445fb
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
import json
import re

from glob import glob

# for i in glob("data/tweet_topic/*.jsonl"):
#     with open(i) as f:
#         data = [json.loads(j) for j in f.readlines()]
#     for d in data:
#         for c in ['text']:
#             d[c] = d[c].replace("{{USERNAME}}", "@user").replace("{{URL}}", "{URL}")
#             for t in re.findall(r'{@[^@^}]*@}', d[c]):
#                 d[c] = d[c].replace(t, t.replace("{@", "@").replace("@}", "").replace(" ", "_"))
#     with open(i, "w") as f:
#         f.write("\n".join([json.dumps(j) for j in data]))

for i in glob("data/tweet_ner7/*.jsonl"):
    with open(i) as f:
        data = [json.loads(j) for j in f.readlines()]
    for d in data:
        d['text'] = d['text'].replace("{{USERNAME}}", "@user").replace("{{URL}}", "{URL}")
        d['text_tokenized'] = [y if y != "{{USERNAME}}" else "@user" for y in d['text_tokenized']]
        d['text_tokenized'] = [y if y != "{{URL}}" else "{URL}" for y in d['text_tokenized']]

        for t in re.findall(r'{@[^@^}]*@}', d['text']):
            t_new = t.replace("{@", "@").replace("@}", "").replace(" ", "_")
            d['text'] = d['text'].replace(t, t_new)
            d['text_tokenized'] = [y if y != t else t_new for y in d['text_tokenized']]
            for e in d['entities']:
                e['entity'] = e['entity'].replace(t, t_new)
                e['entity'] = e['entity'].replace("{{USERNAME}}", "@user").replace("{{URL}}", "{URL}")

    with open(i, "w") as f:
        f.write("\n".join([json.dumps(j) for j in data]))