ttj commited on
Commit
868b7de
·
1 Parent(s): 0d6806f

Saving weights and logs of epoch 1

Browse files
config.json CHANGED
@@ -11,7 +11,7 @@
11
  "intermediate_size": 3072,
12
  "layer_norm_eps": 1e-12,
13
  "max_position_embeddings": 512,
14
- "model_type": "f_bert",
15
  "num_attention_heads": 12,
16
  "num_hidden_layers": 12,
17
  "pad_token_id": 0,
 
11
  "intermediate_size": 3072,
12
  "layer_norm_eps": 1e-12,
13
  "max_position_embeddings": 512,
14
+ "model_type": "bert",
15
  "num_attention_heads": 12,
16
  "num_hidden_layers": 12,
17
  "pad_token_id": 0,
events.out.tfevents.1625285291.b550.5766.3.v2 ADDED
Binary file (25.2 kB). View file
 
flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9b0e97ede96897228cf4f9f8deb40fa7777897177ce66f3fbd4769cf3ec00a60
3
- size 348322811
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea83cb573d80d7f5a8f572e243be5d871997d83097a64ff4fb69e000c5b5ee86
3
+ size 433370483