pere commited on
Commit
d28cf3c
1 Parent(s): 2d4285d

Saving weights and logs of step 10000

Browse files
config.json CHANGED
@@ -20,7 +20,7 @@
20
  "num_hidden_layers": 12,
21
  "pad_token_id": 1,
22
  "position_embedding_type": "absolute",
23
- "torch_dtype": "bloat16",
24
  "transformers_version": "4.15.0.dev0",
25
  "type_vocab_size": 1,
26
  "use_cache": true,
 
20
  "num_hidden_layers": 12,
21
  "pad_token_id": 1,
22
  "position_embedding_type": "absolute",
23
+ "torch_dtype": "bfloat16",
24
  "transformers_version": "4.15.0.dev0",
25
  "type_vocab_size": 1,
26
  "use_cache": true,
events.out.tfevents.1640779990.t1v-n-358ff5d1-w-0.2119930.3.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce8cbb1694cdf68e6511dfe36685860589e85588faaeee1d29b3d86d2c5b5ec0
3
+ size 1470136
flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34022cc8cc708f928034e0ae78a30f59e7249d556955b2fb7cdff3f427cb000a
3
  size 498796983
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e126ffe6605b0ee84fefc76010690220a8469d33af635ef3ca6dcba3fe8b231f
3
  size 498796983
run_step1.sh CHANGED
@@ -22,5 +22,5 @@
22
  --preprocessing_num_workers="64" \
23
  --auth_token="True" \
24
  --static_learning_rate="True" \
25
- --dtype="bloat16" \
26
  --push_to_hub
 
22
  --preprocessing_num_workers="64" \
23
  --auth_token="True" \
24
  --static_learning_rate="True" \
25
+ --dtype="bfloat16" \
26
  --push_to_hub