Add new SentenceTransformer model.
Browse files- .gitattributes +2 -0
- added_tokens.json +1 -1
- config.json +1 -1
- pytorch_model.bin +1 -1
- tokenizer_config.json +1 -1
.gitattributes
CHANGED
@@ -28,3 +28,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
28 |
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
29 |
pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
|
30 |
.git/lfs/objects/32/65/32656706e38c4ffb554cd11524a935b5ed31b6685a661db4d802d062a2e34c3c filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
28 |
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
29 |
pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
|
30 |
.git/lfs/objects/32/65/32656706e38c4ffb554cd11524a935b5ed31b6685a661db4d802d062a2e34c3c filter=lfs diff=lfs merge=lfs -text
|
31 |
+
.git/lfs/objects/06/d1/06d1de8984836f80f4d67ad569562fe9836cfbbd185be6fd908a4a43d545e001 filter=lfs diff=lfs merge=lfs -text
|
32 |
+
.git/lfs/objects/94/cb/94cb37f01aab796c19281c16c74ec08010630bc4b4ec7f84f764bb2ba357b8b9 filter=lfs diff=lfs merge=lfs -text
|
added_tokens.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"<
|
|
|
1 |
+
{"<A>": 250003, "<link>": 250004, "<Q>": 250002}
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "output/
|
3 |
"architectures": [
|
4 |
"XLMRobertaModel"
|
5 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "output/1024/checkpoint-3000",
|
3 |
"architectures": [
|
4 |
"XLMRobertaModel"
|
5 |
],
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1112262449
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dd4269157ebf5f786ed8478fd1339aaa3fd582c24acebe1f56ca8a6c228df407
|
3 |
size 1112262449
|
tokenizer_config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "additional_special_tokens": ["<Q>", "<A>", "<link>"], "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "output/
|
|
|
1 |
+
{"bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "additional_special_tokens": ["<Q>", "<A>", "<link>"], "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "output/1024/checkpoint-3000", "tokenizer_class": "XLMRobertaTokenizer"}
|