Commit
•
72a7d56
1
Parent(s):
e1bb252
Create evalharness.sh
Browse files- evalharness.sh +119 -0
evalharness.sh
ADDED
@@ -0,0 +1,119 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
#!/bin/bash
|
2 |
+
#SBATCH --job-name=run_evalharness-tr11f-6b3-ml
|
3 |
+
#SBATCH --partition=gpu_p5
|
4 |
+
#SBATCH --constraint=a100
|
5 |
+
#SBATCH --nodes=1
|
6 |
+
#SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node!
|
7 |
+
#SBATCH --cpus-per-task=8 # number of cores per tasks
|
8 |
+
#SBATCH --hint=nomultithread # we get physical cores not logical
|
9 |
+
#SBATCH --gres=gpu:1 # number of gpus
|
10 |
+
#SBATCH --time 20:00:00 # maximum execution time (HH:MM:SS)
|
11 |
+
#SBATCH --output=%x-%j.out # output file name
|
12 |
+
#SBATCH --account=ajs@a100
|
13 |
+
|
14 |
+
set -x -e
|
15 |
+
|
16 |
+
source $six_ALL_CCFRWORK/start-py38-pt111
|
17 |
+
|
18 |
+
echo "START TIME: $(date)"
|
19 |
+
|
20 |
+
# a unique identifier for the current eval ideally correspnding to the modelname
|
21 |
+
VARIANT="tr11f-6b3-ml-evalharness"
|
22 |
+
|
23 |
+
|
24 |
+
CHECKPOINT_PATH=$six_ALL_CCFRSCRATCH/checkpoints/tr13f-6B3-ml-t0/checkpoints/tasky/global_step1000
|
25 |
+
MEGATRON_DEEPSPEED_REPO=/gpfsssd/worksf/projects/rech/six/commun/code/eval/Megatron-DeepSpeed
|
26 |
+
export HF_DATASETS_OFFLINE=1
|
27 |
+
export TRANSFORMERS_OFFLINE=1
|
28 |
+
|
29 |
+
export TRANSFORMERS_CACHE=$six_ALL_CCFRWORK/models
|
30 |
+
export HF_DATASETS_CACHE=$six_ALL_CCFRWORK/datasets
|
31 |
+
export HF_MODULES_CACHE=$six_ALL_CCFRWORK/modules
|
32 |
+
export HF_METRICS_CACHE=$six_ALL_CCFRWORK/metrics
|
33 |
+
|
34 |
+
cd $MEGATRON_DEEPSPEED_REPO
|
35 |
+
|
36 |
+
TOKENIZER_NAME_OR_PATH=bigscience-catalogue-data-dev/byte-level-bpe-tokenizer-no-norm-250k-whitespace-and-eos-regex-alpha-v3-dedup-lines-articles
|
37 |
+
|
38 |
+
PP_SIZE=1
|
39 |
+
TP_SIZE=1
|
40 |
+
SEQ_LEN=2048
|
41 |
+
|
42 |
+
# different from the training MICRO_BATCH_SIZE - no optim memory, so can do bigger BS
|
43 |
+
# make as big as it can fit into gpu w/o OOM, but not too close to 100%
|
44 |
+
EVAL_MICRO_BATCH_SIZE=1
|
45 |
+
|
46 |
+
#dummy arguments to make megatron happy.
|
47 |
+
MEGATRON_REQUIRED_ARGS=" \
|
48 |
+
--num-layers -1 \
|
49 |
+
--hidden-size -1 \
|
50 |
+
--num-attention-heads -1 \
|
51 |
+
--seq-length -1 \
|
52 |
+
--max-position-embeddings -1 \
|
53 |
+
"
|
54 |
+
|
55 |
+
|
56 |
+
ZERO_STAGE=0
|
57 |
+
|
58 |
+
config_json="./ds_config.json"
|
59 |
+
|
60 |
+
# Deepspeed figures out GAS dynamically from dynamic GBS via set_train_batch_size()
|
61 |
+
cat <<EOT > $config_json
|
62 |
+
{
|
63 |
+
"train_micro_batch_size_per_gpu": 1,
|
64 |
+
"train_batch_size": 1,
|
65 |
+
"gradient_clipping": 1.0,
|
66 |
+
"zero_optimization": {
|
67 |
+
"stage": $ZERO_STAGE
|
68 |
+
},
|
69 |
+
"bf16": {
|
70 |
+
"enabled": false
|
71 |
+
},
|
72 |
+
"steps_per_print": 2000,
|
73 |
+
"wall_clock_breakdown": false
|
74 |
+
}
|
75 |
+
EOT
|
76 |
+
|
77 |
+
|
78 |
+
CMD="./tasks/eval_harness/evaluate.py \
|
79 |
+
--load $CHECKPOINT_PATH \
|
80 |
+
--results_path $VARIANT-results.json \
|
81 |
+
--tensor-model-parallel-size $TP_SIZE \
|
82 |
+
--pipeline-model-parallel-size $PP_SIZE \
|
83 |
+
--tokenizer-type PretrainedFromHF \
|
84 |
+
--tokenizer-name-or-path $TOKENIZER_NAME_OR_PATH \
|
85 |
+
--micro-batch-size $EVAL_MICRO_BATCH_SIZE \
|
86 |
+
--no-load-optim \
|
87 |
+
--no-load-rng \
|
88 |
+
--eval_fp32 \
|
89 |
+
--inference \
|
90 |
+
--seq-length $SEQ_LEN \
|
91 |
+
--task_list arc_challenge,arc_easy,boolq,copa,headqa,hellaswag,lambada,logiqa,mathqa,mc_taco,mrpc,multirc,openbookqa,piqa,prost,pubmedqa,qnli,qqp,race,rte,sciq,sst,triviaqa,webqs,wic,winogrande,wnli,wsc \
|
92 |
+
--deepspeed \
|
93 |
+
--deepspeed_config ds_config.json \
|
94 |
+
--intermed_results \
|
95 |
+
--adaptive_seq_len \
|
96 |
+
--micro_bs_multiplier 4 \
|
97 |
+
$MEGATRON_REQUIRED_ARGS \
|
98 |
+
"
|
99 |
+
|
100 |
+
GPUS_PER_NODE=1
|
101 |
+
NNODES=$SLURM_NNODES
|
102 |
+
MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1)
|
103 |
+
MASTER_PORT=6000
|
104 |
+
export LAUNCHER="python -u -m torch.distributed.run \
|
105 |
+
--nproc_per_node $GPUS_PER_NODE \
|
106 |
+
--nnodes $NNODES \
|
107 |
+
--rdzv_endpoint $MASTER_ADDR:$MASTER_PORT \
|
108 |
+
--rdzv_backend c10d \
|
109 |
+
--max_restarts 0 \
|
110 |
+
--tee 3 \
|
111 |
+
"
|
112 |
+
|
113 |
+
export CUDA_LAUNCH_BLOCKING=1
|
114 |
+
|
115 |
+
echo $LAUNCHER $CMD
|
116 |
+
|
117 |
+
export PYTHONPATH=$MEGATRON_DEEPSPEED_REPO
|
118 |
+
|
119 |
+
$LAUNCHER $CMD 2>&1 | tee $VARIANT-eval-harness.log
|