pere commited on
Commit
4666884
1 Parent(s): 2aba2ce
batch_lrtest.sh ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ PROJECT_DIR=${HOME}"/models/t5-nynorsk-norbench"
3
+ export PYTHONPATH=${PROJECT_DIR}
4
+
5
+ FINETUNE_STEPS=10000
6
+ EVAL_PREFIX="norbench/eval_lr_translate_"
7
+ MODEL_BUCKET_DIR="gs://pere-north-t5x/finetuned/"
8
+
9
+ CHECKPOINT_LIST=(
10
+ "pere-north-t5x/pretrained_models/base/norwegian_NCC_plus_English_t5x_base/checkpoint_1500000"
11
+ "pere-north-t5x/pretrained_models/base/norwegian_NCC_plus_English_t5x_base/checkpoint_1500000"
12
+ "pere-north-t5x/pretrained_models/base/norwegian_NCC_plus_English_t5x_base/checkpoint_1500000"
13
+ "pere-north-t5x/pretrained_models/base/norwegian_NCC_plus_English_t5x_base/checkpoint_1500000"
14
+ "pere-north-t5x/pretrained_models/base/norwegian_NCC_plus_English_t5x_base/checkpoint_1500000"
15
+ "pere-north-t5x/pretrained_models/base/norwegian_NCC_plus_English_t5x_base/checkpoint_1500000"
16
+ )
17
+
18
+ NAME_LIST=(
19
+ "north_t5_small_NCC"
20
+ "north_t5_base_NCC"
21
+ "north_t5_large_NCC"
22
+ "mt5_small"
23
+ "mt5_base"
24
+ "mt5_large"
25
+ )
26
+
27
+ TASK_LIST=("translate_mt5" "translate_mt5" "translate_mt5" "translate_mt5" "translate_mt5" "translate_mt5")
28
+ LR_LIST=(0.00001 0.0001 0.0005 0.00075 0.001 0.002)
29
+
30
+
31
+ GIN_LIST=(
32
+ "finetune_translate_base_mt5_lr.gin"
33
+ )
34
+
35
+ START_LIST=(1500000 1500000 1500000 1000000 1000000 1000000)
36
+ EXP_LIST=(115 116 117 118 119 120)
37
+
38
+
39
+ VERSION_LIST=("1")
40
+
41
+ index=$(($1 + 0))
42
+
43
+ if [ $index -lt 1 ] || [ $index -gt ${#CHECKPOINT_LIST[@]} ] || [ $# -ne 1 ]; then
44
+ echo "Error: You need to provide the number of the checkpoints below as a parameter."
45
+ for i in "${!CHECKPOINT_LIST[@]}"; do
46
+ echo "$((i+1)). ${CHECKPOINT_LIST[i]}"
47
+ done
48
+
49
+ exit 1
50
+ fi
51
+
52
+ for v in "${VERSION_LIST[@]}"; do
53
+ i=($index-1)
54
+ INITIAL_CHECKPOINT_PATH="gs://${CHECKPOINT_LIST[i]}"
55
+
56
+ TRAIN_STEPS=$((START_LIST[i]+FINETUNE_STEPS))
57
+ GIN_FILE=${GIN_LIST[i]}
58
+ MIXTURE_OR_TASK_NAME=${TASK_LIST[i]}
59
+ LR=${LR_LIST[i]}
60
+ MODEL_DIR="${MODEL_BUCKET_DIR}${EVAL_PREFIX}lr${LR}_exp${EXP_LIST[i]}_${NAME_LIST[i]}_v${v}"
61
+
62
+ command="python3 ../../t5x/t5x/train.py --gin_search_paths=\"./\" --gin.TRAIN_STEPS=${TRAIN_STEPS} --gin.LR=${LR} --gin_file=${GIN_FILE} --gin.INITIAL_CHECKPOINT_PATH=\\\"${INITIAL_CHECKPOINT_PATH}\\\" --gin.MIXTURE_OR_TASK_NAME=\\\"${MIXTURE_OR_TASK_NAME}\\\" --gin.MODEL_DIR=\\\"${MODEL_DIR}\\\""
63
+ echo "${command}"
64
+ # Uncomment the next line to run the command:
65
+ eval "${command}"
66
+ done
67
+
68
+ #python3 ../../t5x/t5x/train.py --gin_search_paths="./" --gin.TRAIN_STEPS=${TRAIN_STEPS} --gin_file="finetune_translate_base.gin" --gin.INITIAL_CHECKPOINT_PATH=${INITIAL_CHECKPOINT_PATH} --gin.MIXTURE_OR_TASK_NAME=\"translate_long\" --gin.MODEL_DIR=\"gs://nb-t5x-us-central2/finetuned/nynorsk_NCC_base_v1\" &&
69
+ #python3 ../../t5x/t5x/train.py --gin_search_paths="./" --gin.TRAIN_STEPS=${TRAIN_STEPS} --gin_file="finetune_translate_base.gin" --gin.INITIAL_CHECKPOINT_PATH=${INITIAL_CHECKPOINT_PATH} --gin.MIXTURE_OR_TASK_NAME=\"translate_long\" --gin.MODEL_DIR=\"gs://nb-t5x-us-central2/finetuned/nynorsk_NCC_base_v2\" &&
70
+ #python3 ../../t5x/t5x/train.py --gin_search_paths="./" --gin.TRAIN_STEPS=${TRAIN_STEPS} --gin_file="finetune_translate_base.gin" --gin.INITIAL_CHECKPOINT_PATH=${INITIAL_CHECKPOINT_PATH} --gin.MIXTURE_OR_TASK_NAME=\"translate_long\" --gin.MODEL_DIR=\"gs://nb-t5x-us-central2/finetuned/nynorsk_NCC_base_v3\" &&
71
+ #python3 ../../t5x/t5x/train.py --gin_search_paths="./" --gin.TRAIN_STEPS=${TRAIN_STEPS} --gin_file="finetune_translate_base.gin" --gin.INITIAL_CHECKPOINT_PATH=${INITIAL_CHECKPOINT_PATH} --gin.MIXTURE_OR_TASK_NAME=\"translate_long\" --gin.MODEL_DIR=\"gs://nb-t5x-us-central2/finetuned/nynorsk_NCC_base_v4\" &&
72
+ #python3 ../../t5x/t5x/train.py --gin_search_paths="./" --gin.TRAIN_STEPS=${TRAIN_STEPS} --gin_file="finetune_translate_base.gin" --gin.INITIAL_CHECKPOINT_PATH=${INITIAL_CHECKPOINT_PATH} --gin.MIXTURE_OR_TASK_NAME=\"translate_long\" --gin.MODEL_DIR=\"gs://nb-t5x-us-central2/finetuned/nynorsk_NCC_base_v5\"
73
+
finetune_translate_base_mt5_lr.gin ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from __gin__ import dynamic_registration
2
+ import tasks
3
+ import seqio
4
+
5
+ import __main__ as train_script
6
+ from t5.data import mixtures
7
+ from t5x import models
8
+ from t5x import partitioning
9
+ from t5x import utils
10
+
11
+ include 't5x/examples/t5/mt5/base.gin'
12
+ include "t5x/configs/runs/finetune.gin"
13
+
14
+ MIXTURE_OR_TASK_NAME = %gin.REQUIRED
15
+ TASK_FEATURE_LENGTHS = {"inputs": 256, "targets": 256}
16
+ INITIAL_CHECKPOINT_PATH = %gin.REQUIRED
17
+ LR = %gin.REQUIRED
18
+ TRAIN_STEPS = %gin.REQUIRED # 1000000 pre-trained steps + 10000 fine-tuning steps.
19
+ USE_CACHED_TASKS = False
20
+ DROPOUT_RATE = 0.1
21
+ RANDOM_SEED = 0
22
+
23
+ #Fixing a small error
24
+ infer_eval/utils.DatasetConfig:
25
+ task_feature_lengths = %TASK_FEATURE_LENGTHS
26
+
27
+ #Saving every 500 steps
28
+ utils.SaveCheckpointConfig:
29
+ period = 1000
30
+ keep = 1 # number of checkpoints to keep
31
+
32
+ utils.create_learning_rate_scheduler:
33
+ factors = 'constant'
34
+ base_learning_rate = %LR
35
+ warmup_steps = 1000
36
+
37
+ # Might have to ba changed based on architecture
38
+ # partitioning.PjitPartitioner.num_partitions = 1
39
+
40
+