HoangHa commited on
Commit
e5c7b1a
1 Parent(s): 99441a7

Saving weights and logs of epoch 4

Browse files
epoch_4/config.json CHANGED
@@ -12,7 +12,7 @@
12
  "initializer_range": 0.02,
13
  "intermediate_size": 3072,
14
  "layer_norm_eps": 1e-12,
15
- "max_position_embeddings": 130,
16
  "model_type": "roberta",
17
  "num_attention_heads": 12,
18
  "num_hidden_layers": 12,
 
12
  "initializer_range": 0.02,
13
  "intermediate_size": 3072,
14
  "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 66,
16
  "model_type": "roberta",
17
  "num_attention_heads": 12,
18
  "num_hidden_layers": 12,
epoch_4/flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2e2f6f0ded38c85f8d8bbba83265b0ef833dbc70199647d649e8127425dd358a
3
- size 352639294
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c95f351e6f007407729a2141ccc2a469debb9265bfe0da8ff72724b9a183f40
3
+ size 352442685