|
import glob, pandas, sys, os, gzip, json, time, lzma |
|
|
|
try: fast_mode = (sys.argv[2] == "fast") |
|
except: fast_mode = False |
|
|
|
offset = 10 |
|
index = int(sys.argv[1]) * offset |
|
|
|
dumps = glob.glob("CC-MAIN-*") |
|
for dump in dumps: |
|
all_buckets = glob.glob(f"{dump}/CC-MAIN-*") |
|
buckets = all_buckets[index : index + offset] |
|
print("\n\n- - - - -\n", dump, "has", len(all_buckets), flush=True) |
|
print(fast_mode, index, offset, buckets, flush=True) |
|
for bucket in buckets: |
|
files = glob.glob(f"{bucket}/*.parquet") |
|
bucket = bucket.replace(f"{dump}/", f"{dump}_") |
|
bucket = bucket.replace("CC-MAIN-", "") |
|
output_file = f"jsonl/{bucket}.jsonl.lzma" |
|
print(output_file, flush=True) |
|
|
|
added_file_ids = set() |
|
if os.path.exists(output_file): |
|
ti_mcur = os.path.getmtime(output_file) |
|
ti_mmax = 0 |
|
remains = [] |
|
for file in files: |
|
ti_m = os.path.getmtime(file) |
|
if ti_m > ti_mmax: ti_mmax = ti_m |
|
if fast_mode and ti_m > ti_mcur: |
|
remains.append(file) |
|
|
|
if ti_mcur > ti_mmax: continue |
|
|
|
if fast_mode: |
|
files = remains |
|
else: |
|
skip_token = '"id": "000000000."' |
|
with lzma.open(output_file, "rt") as fin: |
|
for count, line in enumerate(fin): |
|
if skip_token not in line: |
|
file_id = line.split('"id": "')[1].split(".")[0] |
|
skip_token = f'"id": "{file_id}.' |
|
assert file_id not in added_file_ids |
|
added_file_ids.add(file_id) |
|
|
|
with lzma.open(output_file, "at") as fout: |
|
for file in files: |
|
file_id = file.split("/")[-1].replace(".parquet", "") |
|
if file_id in added_file_ids: continue |
|
print(f"Adding {file_id} to {output_file}...", flush=True) |
|
df = pandas.read_parquet(file) |
|
for line_count, row in df.iterrows(): |
|
idd = f"{file_id}.{line_count}" |
|
ss = json.dumps({"text": row[0], "id": idd}, ensure_ascii=False) |
|
fout.write(ss + "\n") |
|
|
|
print("fast_mode", fast_mode) |
|
|