bigjoedata commited on
Commit
6721bba
1 Parent(s): 8f9fc7a

Retrained and cleaned model

Browse files
config.json CHANGED
@@ -25,6 +25,7 @@
25
  "summary_proj_to_labels": true,
26
  "summary_type": "cls_index",
27
  "summary_use_proj": true,
 
28
  "use_cache": true,
29
  "vocab_size": 50257
30
  }
 
25
  "summary_proj_to_labels": true,
26
  "summary_type": "cls_index",
27
  "summary_use_proj": true,
28
+ "transformers_version": "4.2.2",
29
  "use_cache": true,
30
  "vocab_size": 50257
31
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc2a4ed0e553d18709a7482fa3b59e5b4324c7916036e57be8664763eb2c4ef5
3
- size 510404221
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfe4adb9e2b2bb1d5fd5a0432f925b871e0de2473568bc907e5a7a305d3cf039
3
+ size 510404194
special_tokens_map.json CHANGED
@@ -1 +1 @@
1
- {"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "pad_token": "<|endoftext|>"}
 
1
+ {"bos_token": {"content": "<|startoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "pad_token": "<|endoftext|>"}
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"errors": "replace", "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "pad_token": "<|endoftext|>"}
 
1
+ {"errors": "replace", "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|startoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "pad_token": "<|endoftext|>"}