| import argparse |
| from collections import namedtuple |
| import os |
|
|
| DATADIR = "/path/to/train_data" |
| DEDUP_FROM_DIR = "/path/to/eval/data" |
| OUTPUT_DIR = "/path/to/output/data" |
|
|
|
|
| def main(args): |
| languages = set() |
| for language_directory in os.listdir(DATADIR): |
| if "_" in language_directory: |
| src, tgt = language_directory.split("_") |
| languages.add(LanguagePair(src=src, tgt=tgt)) |
|
|
| data = existing_data() |
| train_languages = sorted(languages) |
| for language_pair in train_languages[args.start_index:args.start_index + args.size]: |
| print(language_pair) |
| dedup(language_pair, data) |
|
|
|
|
| LanguagePair = namedtuple("LanguagePair", ["src", "tgt"]) |
|
|
|
|
| def existing_data(): |
| data = set() |
| for file in os.listdir(DEDUP_FROM_DIR): |
| with open(os.path.join(DEDUP_FROM_DIR, file)) as f: |
| data |= set(f.readlines()) |
| return data |
| |
| def dedup(language_pair, data, verbose=True, output=True): |
| train_filenames = LanguagePair( |
| src=f"{DATADIR}/{language_pair.src}_{language_pair.tgt}/train.{language_pair.src}", |
| tgt=f"{DATADIR}/{language_pair.src}_{language_pair.tgt}/train.{language_pair.tgt}", |
| ) |
|
|
| output_filenames = LanguagePair( |
| src=f"{OUTPUT_DIR}/train.dedup.{language_pair.src}-{language_pair.tgt}.{language_pair.src}", |
| tgt=f"{OUTPUT_DIR}/train.dedup.{language_pair.src}-{language_pair.tgt}.{language_pair.tgt}" |
| ) |
|
|
| |
| if (os.path.exists(output_filenames.src) and |
| os.path.exists(output_filenames.tgt)): |
| if verbose: |
| print(f"{language_pair.src}-{language_pair.tgt} already done.") |
| return |
|
|
| if verbose: |
| print(f"{language_pair.src}-{language_pair.tgt} ready, will check dups.") |
|
|
| |
| if not output: |
| return |
|
|
| if os.path.exists(train_filenames.src) and os.path.exists(train_filenames.tgt): |
| with open(train_filenames.src) as f: |
| train_source = f.readlines() |
|
|
| with open(train_filenames.tgt) as f: |
| train_target = f.readlines() |
|
|
| |
| new_train_source = [] |
| new_train_target = [] |
| for i, train_line in enumerate(train_source): |
| if train_line not in data and train_target[i] not in data: |
| new_train_source.append(train_line) |
| new_train_target.append(train_target[i]) |
|
|
| assert len(train_source) == len(train_target) |
| assert len(new_train_source) == len(new_train_target) |
| assert len(new_train_source) <= len(train_source) |
|
|
| with open(output_filenames.src, "w") as o: |
| for line in new_train_source: |
| o.write(line) |
|
|
| with open(output_filenames.tgt, "w") as o: |
| for line in new_train_target: |
| o.write(line) |
|
|
|
|
| if __name__ == '__main__': |
| parser = argparse.ArgumentParser() |
| parser.add_argument("-s", "--start-index", required=True, type=int) |
| parser.add_argument("-n", "--size", required=True, type=int) |
| main(parser.parse_args()) |
|
|