pere commited on
Commit
9b3f1e5
1 Parent(s): 80363ae

Saving weights and logs of step 5000

Browse files
config.json CHANGED
@@ -21,7 +21,7 @@
21
  "pad_token_id": 1,
22
  "position_embedding_type": "absolute",
23
  "torch_dtype": "float32",
24
- "transformers_version": "4.13.0.dev0",
25
  "type_vocab_size": 1,
26
  "use_cache": true,
27
  "vocab_size": 50265
 
21
  "pad_token_id": 1,
22
  "position_embedding_type": "absolute",
23
  "torch_dtype": "float32",
24
+ "transformers_version": "4.14.0.dev0",
25
  "type_vocab_size": 1,
26
  "use_cache": true,
27
  "vocab_size": 50265
events.out.tfevents.1639565282.t1v-n-4e27a527-w-0.1048792.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc85c616480f85dfbb4573c7f5e287a0efacd1b36a9c381e854194bf17041f2b
3
+ size 227952
flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e02d522d7ecb7b42196b90da857e13a255d9e302e29aced79963d8bb217627b1
3
  size 498796983
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25132b2566e378641f523dbb48814ea6ad648023271faa8dafd35147c247a880
3
  size 498796983
run_test.sh ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ./run_mlm_flax.py \
2
+ --output_dir="./" \
3
+ --model_type="roberta" \
4
+ --config_name="./" \
5
+ --tokenizer_name="./" \
6
+ --max_seq_length="128" \
7
+ --weight_decay="0.01" \
8
+ --train_file="/mnt/disks/flaxdisk/divided_corpus/train-shard-0001-of-0001.json" \
9
+ --validation_file="/mnt/disks/flaxdisk/divided_corpus/validation-shard-0001-of-0001.json" \
10
+ --per_device_train_batch_size="128" \
11
+ --per_device_eval_batch_size="128" \
12
+ --learning_rate="3e-4" \
13
+ --warmup_steps="5000" \
14
+ --overwrite_output_dir \
15
+ --num_train_epochs="29" \
16
+ --adam_beta1="0.9" \
17
+ --adam_beta2="0.98" \
18
+ --logging_steps="5000" \
19
+ --save_steps="5000" \
20
+ --eval_steps="5000" \
21
+ --push_to_hub