sea_wiki / count_data_stats.py
sabilmakbar's picture
Update the data w/ compressed files to speed-up download process and reduce download size (#6)
5fdddf9
raw
history blame
1.68 kB
import multiprocessing as mp
import numpy as np
from datasets import load_dataset
import tiktoken
def num_tokens_from_string(string: str):
"""Returns the number of tokens in a text string."""
num_tokens = len(encoding.encode(string))
return num_tokens
def cnt_token_in_hf_wiki_dset(data):
data["token_cnt"] = num_tokens_from_string(data["text"])
return data
if __name__ == "__main__":
#this will refer to its local version of dataset loader script, not the HF repo ones
dataset = load_dataset("sea_wiki.py")
encoding = tiktoken.encoding_for_model('gpt-4')
stat_dict = {}
for split, dset in dataset.items():
dset_text = dset.select_columns(['text'])
print(f"Counting total token in split lang: {split}")
dset_text = dset_text.map(cnt_token_in_hf_wiki_dset, num_proc=max(mp.cpu_count()-2,1))
token_data = list(dset_text["token_cnt"])
total_token = sum(token_data)
avg_token = sum(token_data)/len(token_data)
min_token = min(token_data)
max_token = max(token_data)
deciles = np.percentile(token_data, np.arange(10, 100, 10)).tolist()
stat_dict[split] = {"total": total_token, "avg": avg_token, "min": min_token, "max": max_token, "deciles": deciles}
# for markdown table format
print("| Dataset Lang Code | Total Token | Avg Token per Article | Min Token | Max Token | Token Deciles List |")
print("| :---: | ---: | ---: | ---: | ---: | :--- |")
for key, data in stat_dict.items():
print(f"| {key} | {data['total']:,} | {data['avg']:,} | {data['min']:,} | {data['max']:,} | {[round(num,2) for num in data['deciles']]} |")