birgermoell commited on
Commit
5754438
1 Parent(s): a27c158

Updated model

Browse files
config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "_name_or_path": "./",
3
  "architectures": [
4
- "RobertaForMaskedLM"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
1
  {
2
+ "_name_or_path": ".",
3
  "architectures": [
4
+ "RobertaModel"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
events.out.tfevents.1625852361.t1v-n-98937c84-w-0.556819.3.v2 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:39b9519dfbfe40c6a2db63f24069fef2fdff05fb1e5e9f6b55a0e54e4b376c9c
3
- size 19566987
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37f22c9592894928e6c421ca6e9e1a1adbc10159bd7a6e76e0958a95259ca8e4
3
+ size 19714992
merges.txt ADDED
The diff for this file is too large to render. See raw diff
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c4ae452e4835ee0e3d42b4f6bb1acd8c272d4ac34fa7e2e2ad79e2bb0a8c94d4
3
- size 498858859
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6979d65b6e4fc10b6d8771c457b2ef5e9714d9620c9e5fd60e7a1b6a92e0190
3
+ size 498661169
save_updated.py CHANGED
@@ -2,5 +2,5 @@
2
  from transformers import AutoTokenizer, RobertaModel
3
  model = RobertaModel.from_pretrained('.',from_flax=True)
4
  tokenizer = AutoTokenizer.from_pretrained(".")
5
- tokenizer.save_pretrained('exported_pytorch_model')
6
- model.save_pretrained('exported_pytorch_model')
2
  from transformers import AutoTokenizer, RobertaModel
3
  model = RobertaModel.from_pretrained('.',from_flax=True)
4
  tokenizer = AutoTokenizer.from_pretrained(".")
5
+ tokenizer.save_pretrained('.')
6
+ model.save_pretrained('.')
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false}}
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
1
+ {"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": "<mask>", "special_tokens_map_file": null, "name_or_path": ".", "tokenizer_class": "RobertaTokenizer"}
vocab.json ADDED
The diff for this file is too large to render. See raw diff