aapot commited on
Commit
9b8eb3e
1 Parent(s): e52e3c0

Saving weights and logs of step 10000

Browse files
events.out.tfevents.1640023857.t1v-n-8eba1090-w-0.1757053.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dff3582a3f326fc45154024684d65c41df3357136ac4ff101cfb162d6ab4779
3
+ size 1471447
flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:437d81f676318fce198e645188c5a20180f19278afac4b5b9eec8766214da7c6
3
  size 1421662309
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc5b3ca96a41ec669cb35943c8c732e01b6a98c0da4c9d79947da7eab843a2cc
3
  size 1421662309
start_train.sh CHANGED
@@ -11,8 +11,8 @@ python3 run_mlm_flax.py \
11
  --max_seq_length="512" \
12
  --pad_to_max_length \
13
  --preprocessing_num_workers="64" \
14
- --per_device_train_batch_size="16" \
15
- --per_device_eval_batch_size="16" \
16
  --adam_beta1="0.9" \
17
  --adam_beta2="0.98" \
18
  --adam_epsilon="1e-6" \
 
11
  --max_seq_length="512" \
12
  --pad_to_max_length \
13
  --preprocessing_num_workers="64" \
14
+ --per_device_train_batch_size="8" \
15
+ --per_device_eval_batch_size="8" \
16
  --adam_beta1="0.9" \
17
  --adam_beta2="0.98" \
18
  --adam_epsilon="1e-6" \