File size: 639 Bytes
2095da4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
export RUN_NAME=single_latent

# TODO update to not use tokenizer, instead use gpt2 one
./venv/bin/python train.py \
--t5_model_name_or_path="t5-base" \
--output_dir="output/${RUN_NAME}" \
--overwrite_output_dir \
--do_train \
--n_latent_tokens 1 \
--latent_token_size 32 \
--save_steps="2000" \
--block_size="128" \
--per_device_train_batch_size="100" \
--train_file="INVALID.txt" \
--overwrite_output_dir \
--num_train_epochs="1" \

# 200 batch size, 128 sequence len: ? (breaks)
# 100 batch size, 128 sequence len: 252:38:58
# 10 batch size, 128 sequence len: 281:32:53

# Got ~12 hours to train, want 3 saves so one save every 4 hours