pere commited on
Commit
ea0f374
1 Parent(s): c5f2b61

Saving weights and logs of step 98000

Browse files
events.out.tfevents.1642861064.t1v-n-ccbf3e94-w-0.1931366.3.v2 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f281c944abcdd8896060c4beb524381c7d3d59c5a79e2e7cb23cc24dd3347c72
3
- size 14149680
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a0605763614bc69c42e61abda560f863a18a7ca7f82e0006bb5d3c363b76f15
3
+ size 14298892
flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:45e990a0abb92436b291b57126de67ad58260ea2abf6a0bb790d2193aaa333c3
3
  size 498796983
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdc03e44baf814a867d08e4927379a032a9877960dc3d81604ca46522a2cafe4
3
  size 498796983
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": "<mask>", "special_tokens_map_file": null, "name_or_path": ".", "tokenizer_class": "RobertaTokenizer"}
 
1
+ {"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": "<mask>", "trim_offsets": true, "special_tokens_map_file": null, "name_or_path": "NbAiLab/nb-roberta-base", "tokenizer_class": "RobertaTokenizer"}