sgugger commited on
Commit
8683390
1 Parent(s): 41c9a96
Files changed (3) hide show
  1. pytorch_model.bin +1 -1
  2. tokenizer_config.json +1 -1
  3. training_args.bin +1 -1
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4b7108ad8792e0fcac484f8e45a18c3c9ba7d3484aea9cfea1da0e8e834b23e4
3
  size 334055467
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d10355e0f521f622150077071103b5dc4a8297160c45bf18dad4ac0e584906e4
3
  size 334055467
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false, "__type": "AddedToken"}, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "esberto-small", "tokenizer_class": "RobertaTokenizer"}
 
1
+ {"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false, "__type": "AddedToken"}, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "sgugger/esberto-small", "tokenizer_class": "RobertaTokenizer"}
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8028607595fca8bcbfcca8e8d4d9c5ef3db9d4fe840d18c95fc3a6c3e0b94f5b
3
  size 2607
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6d443b6e6b6237eafb9079a6b7d5a880b45611e76f62c1dea226fc0df80aa82
3
  size 2607