Datasets:

Languages:
English
ArXiv:
Tags:
License:
esconv / process.py
chujiezheng's picture
Upload 5 files
32c54c1
import json
import tqdm
import numpy as np
import multiprocessing as mp
import random
from collections import Counter
random.seed(13)
def _norm(x):
return ' '.join(x.strip().split())
strategies = json.load(open('./strategy.json'))
strategies = [e[1:-1] for e in strategies]
strat2id = {strat: i for i, strat in enumerate(strategies)}
original = json.load(open('./ESConv.json'))
def process_data(d):
dial = []
for uttr in d['dialog']:
text = _norm(uttr['content'])
role = uttr['speaker']
if role == 'seeker':
dial.append({
'text': text,
'speaker': 'usr',
})
else:
dial.append({
'text': text,
'speaker': 'sys',
'strategy': uttr['annotation']['strategy'],
})
d['dialog'] = dial
return d
data = []
for e in map(process_data, tqdm.tqdm(original, total=len(original))):
data.append(e)
emotions = Counter([e['emotion_type'] for e in data])
problems = Counter([e['problem_type'] for e in data])
print('emotion', emotions)
print('problem', problems)
random.shuffle(data)
dev_size = int(0.15 * len(data))
test_size = int(0.15 * len(data))
valid = data[:dev_size]
test = data[dev_size: dev_size + test_size]
train = data[dev_size + test_size:]
print('train', len(train))
with open('./train.txt', 'w') as f:
for e in train:
f.write(json.dumps(e) + '\n')
with open('./sample.json', 'w') as f:
json.dump(train[:10], f, ensure_ascii=False, indent=2)
print('valid', len(valid))
with open('./valid.txt', 'w') as f:
for e in valid:
f.write(json.dumps(e) + '\n')
print('test', len(test))
with open('./test.txt', 'w') as f:
for e in test:
f.write(json.dumps(e) + '\n')