| import json |
| import tqdm |
| import numpy as np |
| import multiprocessing as mp |
| import random |
| from collections import Counter |
| random.seed(13) |
|
|
|
|
| def _norm(x): |
| return ' '.join(x.strip().split()) |
|
|
|
|
| strategies = json.load(open('./strategy.json')) |
| strategies = [e[1:-1] for e in strategies] |
| strat2id = {strat: i for i, strat in enumerate(strategies)} |
| original = json.load(open('./ESConv.json')) |
|
|
| def process_data(d): |
| dial = [] |
| for uttr in d['dialog']: |
| text = _norm(uttr['content']) |
| role = uttr['speaker'] |
| if role == 'seeker': |
| dial.append({ |
| 'text': text, |
| 'speaker': 'usr', |
| }) |
| else: |
| dial.append({ |
| 'text': text, |
| 'speaker': 'sys', |
| 'strategy': uttr['annotation']['strategy'], |
| }) |
| d['dialog'] = dial |
| return d |
|
|
| data = [] |
|
|
| for e in map(process_data, tqdm.tqdm(original, total=len(original))): |
| data.append(e) |
|
|
| emotions = Counter([e['emotion_type'] for e in data]) |
| problems = Counter([e['problem_type'] for e in data]) |
| print('emotion', emotions) |
| print('problem', problems) |
|
|
|
|
| random.shuffle(data) |
| dev_size = int(0.15 * len(data)) |
| test_size = int(0.15 * len(data)) |
| valid = data[:dev_size] |
| test = data[dev_size: dev_size + test_size] |
| train = data[dev_size + test_size:] |
|
|
| print('train', len(train)) |
| with open('./train.txt', 'w') as f: |
| for e in train: |
| f.write(json.dumps(e) + '\n') |
| with open('./sample.json', 'w') as f: |
| json.dump(train[:10], f, ensure_ascii=False, indent=2) |
|
|
| print('valid', len(valid)) |
| with open('./valid.txt', 'w') as f: |
| for e in valid: |
| f.write(json.dumps(e) + '\n') |
|
|
| print('test', len(test)) |
| with open('./test.txt', 'w') as f: |
| for e in test: |
| f.write(json.dumps(e) + '\n') |
|
|