eduagarcia commited on
Commit
f7d589f
1 Parent(s): e9d7237

rollback to 62500 steps

Browse files
Files changed (2) hide show
  1. pytorch_model.bin +1 -1
  2. tokenizer_config.json +1 -1
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5bad26ca497d9910adb9a957fa59cfa31e1a02b9d45138bd096550ba066671f0
3
  size 498843795
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b725fc79a9343628d2aa1c2ea17fe85e8114d0ea915a331b23773117e8a2257
3
  size 498843795
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"errors": "replace", "bos_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "sep_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "cls_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "pad_token": {"content": "<pad>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "/workspace/models/roberta_pt/roberta-pt/mestrado/ptcrawl_plus_legal_base_v1_7/checkpoint_last", "tokenizer_class": "RobertaTokenizer"}
 
1
+ {"errors": "replace", "bos_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "sep_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "cls_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "pad_token": {"content": "<pad>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "/workspace/models/roberta_pt/roberta-pt/mestrado/ptcrawl_plus_legal_base_v3_5/checkpoint_last", "tokenizer_class": "RobertaTokenizer"}