salti commited on
Commit
d2d5ea1
1 Parent(s): a1c5968

Saving weights and logs of step 1000

Browse files
config.json CHANGED
@@ -23,5 +23,5 @@
23
  "tie_word_embeddings": false,
24
  "transformers_version": "4.9.0.dev0",
25
  "use_cache": true,
26
- "vocab_size": 64003
27
  }
 
23
  "tie_word_embeddings": false,
24
  "transformers_version": "4.9.0.dev0",
25
  "use_cache": true,
26
+ "vocab_size": 64103
27
  }
events.out.tfevents.1626415996.t1v-n-a5dc64a9-w-0.41138.3.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3743159f246f5f2f4b63c20b39c7323c9214f1c0cfc428d8c43b7e5f48c990f2
3
+ size 147756
flax_model.msgpack ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:437f7cb4dbacd6462c43e4b7b365215873cd90d352cbd4a55c8ff7408d2cb6ae
3
+ size 438822439
run-t5v1_1-small.sh DELETED
@@ -1,27 +0,0 @@
1
- export model_dir=arabic-t5-small
2
- export train_batch_size=48
3
- export eval_batch_size=48
4
-
5
- python ./run_t5_mlm_flax.py \
6
- --model_type t5 \
7
- --config_name ${model_dir} \
8
- --tokenizer_name ${model_dir} \
9
- --use_fast_tokenizer True \
10
- --dtype float32 \
11
- --max_seq_length 512 \
12
- --preprocessing_num_workers 96 \
13
- --output_dir ${model_dir} \
14
- --overwrite_output_dir True \
15
- --do_train \
16
- --per_device_train_batch_size ${train_batch_size} \
17
- --per_device_eval_batch_size ${eval_batch_size} \
18
- --learning_rate 1e-2 \
19
- --num_train_epochs 1 \
20
- --logging_steps 100 \
21
- --eval_steps 2500 \
22
- --save_steps 10000 \
23
- --seed 12 \
24
- --adafactor True \
25
- --push_to_hub \
26
- --cache_dir ./training_cache \
27
- --save_total_limit 5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run-t5v1_1-small.sh ADDED
@@ -0,0 +1 @@
 
 
1
+ ../run-t5v1_1-small.sh