Saving weights and logs of step 10000
Browse files
config.json
CHANGED
@@ -21,7 +21,7 @@
|
|
21 |
"pad_token_id": 1,
|
22 |
"position_embedding_type": "absolute",
|
23 |
"torch_dtype": "bfloat16",
|
24 |
-
"transformers_version": "4.
|
25 |
"type_vocab_size": 1,
|
26 |
"use_cache": true,
|
27 |
"vocab_size": 50265
|
|
|
21 |
"pad_token_id": 1,
|
22 |
"position_embedding_type": "absolute",
|
23 |
"torch_dtype": "bfloat16",
|
24 |
+
"transformers_version": "4.16.0.dev0",
|
25 |
"type_vocab_size": 1,
|
26 |
"use_cache": true,
|
27 |
"vocab_size": 50265
|
events.out.tfevents.1640973969.t1v-n-79f0077b-w-0.424445.0.v2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b1ab934c2538a82880a5534c4155647d1f21a700b501f298ac96abc4a2f24eec
|
3 |
+
size 1470136
|
flax_model.msgpack
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6acf96fdb00903d5dcb970105dd0ff140c34be5727c15e9012fcde146dc21993
|
3 |
+
size 498796983
|