Datasets:

Tasks:
Other
Multilinguality:
multilingual
Size Categories:
100M<n<1B
ArXiv:
Tags:
License:
xP3mt / to_meg.slurm
Muennighoff's picture
Add files
60bc123
#!/bin/bash
#SBATCH --job-name=xp3mt # job name
#SBATCH --ntasks=1 # number of MP tasks
#SBATCH --nodes=1
#SBATCH --cpus-per-task=40 # number of cores per tasks
#SBATCH --hint=nomultithread # we get physical cores not logical
#SBATCH --time=10:00:00 # maximum execution time (HH:MM:SS)
#SBATCH --output=%x-%j.out # output file name
#SBATCH --account=ajs@cpu
#SBATCH --partition=cpu_p1
#SBATCH --qos=qos_cpu-t3
set -x -e
source $six_ALL_CCFRWORK/start-tr13f-6B3-ml-t0
export HF_DATASETS_OFFLINE=1
export TRANSFORMERS_OFFLINE=1
MEGATRON_DEEPSPEED_REPO=$six_ALL_CCFRWORK/code/tr13f-6B3-ml-t0/Megatron-DeepSpeed
TOKENIZER_PATH="bigscience/tokenizer"
#ar bn en es fr gu hi id ig mr ne pa prep.py pt rn sw ta te to_meg.slurm ur vi yo zh
LANGS=(
ar
bn
es
fr
gu
hi
id
ig
mr
ne
pa
pt
sw
ta
te
ur
vi
yo
zh
)
DATA_PATH=/gpfswork/rech/six/commun/bigscience-training/jsonls/xP3mt
for val in {0..20}; do
LANG=${LANGS[$val]}
cd $DATA_PATH/$LANG
# Merge
cat *.jsonl > merged_dups_$LANG.jsonl
# Drop duplicates (~1G / 37G for en) + Shuffle
sort -u merged_dups_$LANG.jsonl | shuf > merged_$LANG.jsonl
OUTPUT=/gpfswork/rech/six/commun/bigscience-training/xp3mt/xp3_$LANG
cd $MEGATRON_DEEPSPEED_REPO
#python tools/preprocess_data.py \
# --input $DATA_PATH/$LANG/merged_$LANG.jsonl \
# --output-prefix $OUTPUT \
# --dataset-impl mmap \
# --json-key inputs \
# --tokenizer-type PretrainedFromHF \
# --tokenizer-name-or-path $TOKENIZER_PATH \
# --workers 35
#python tools/preprocess_data.py \
# --input $DATA_PATH/$LANG/merged_$LANG.jsonl \
# --output-prefix $OUTPUT \
# --dataset-impl mmap \
# --json-key targets \
# --tokenizer-type PretrainedFromHF \
# --tokenizer-name-or-path $TOKENIZER_PATH \
# --append-eod \
# --prepend-space \
# --workers 35
done