LysandreJik commited on
Commit
7171116
1 Parent(s): 24c54aa
config.json CHANGED
@@ -27,8 +27,8 @@
27
  "num_random_blocks": 3,
28
  "pad_token_id": 0,
29
  "scale_embedding": true,
30
- "transformers_version": "4.10.0.dev0",
31
  "use_bias": false,
32
  "use_cache": true,
33
- "vocab_size": 99
34
  }
27
  "num_random_blocks": 3,
28
  "pad_token_id": 0,
29
  "scale_embedding": true,
30
+ "transformers_version": "4.11.0.dev0",
31
  "use_bias": false,
32
  "use_cache": true,
33
+ "vocab_size": 1007
34
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f2a61c6e573276d043c6ffb6780d1e66345b9f64cda4754c165784e7b1927ff
3
- size 474983
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a861dffdbbbe883758683721b38f95c89580972ea2254a4f8439070128cd2dc4
3
+ size 711143
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
1
+ {"bos_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "sep_token": {"content": "[SEP]", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "pad_token": {"content": "<pad>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "cls_token": {"content": "[CLS]", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "mask_token": {"content": "[MASK]", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true}}
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
1
+ {"pad_token": {"content": "<pad>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "mask_token": {"content": "[MASK]", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "mask_token_sent": null, "offset": 0, "additional_special_tokens": [], "bos_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "sep_token": {"content": "[SEP]", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "cls_token": {"content": "[CLS]", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "special_tokens_map_file": "/home/patrick/.cache/huggingface/transformers/b548e984b09823ed5cea0e622c0ec194a7e07b3c5d2e6b48a7cde84bd179cc7c.4eda581f816a0a941629106f0338c957910ce4839ecf7e3e743bb79523bf7249", "name_or_path": "google/bigbird-pegasus-large-arxiv", "model_max_length": 4096, "tokenizer_class": "PegasusTokenizer"}