yuji96 commited on
Commit
78b215a
1 Parent(s): 8ca9155

Upload 12 files

Browse files
config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.2.2",
4
+ "transformers": "4.28.1",
5
+ "pytorch": "2.0.1+cu118"
6
+ }
7
+ }
dev-metrics.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "best-epoch": 0,
3
+ "best-step": 128,
4
+ "best-dev": 83.61539847191834
5
+ }
log.csv ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,step,loss,sts-dev
2
+ 0,0,inf,51.375121585435735
3
+ 0,32,5.8056640625,78.35678255256113
4
+ 0,64,3.11572265625,82.6244856578063
5
+ 0,96,2.42578125,83.35358238047021
6
+ 0,128,2.176513671875,83.61539847191834
7
+ 0,160,2.0302734375,81.87700855632563
8
+ 0,192,1.875732421875,79.53856626810594
9
+ 0,224,1.843994140625,77.77162870199022
10
+ 0,256,1.773193359375,76.12193564966829
11
+ 1,288,1.690185546875,75.82749644399787
12
+ 1,320,1.600341796875,77.40347138209835
13
+ 1,352,1.56298828125,74.4370737294954
14
+ 1,384,1.570556640625,76.08974073722088
15
+ 1,416,1.534912109375,73.89938923579099
16
+ 1,448,1.55078125,70.69398516957513
17
+ 1,480,1.509521484375,73.33702686894722
18
+ 1,512,1.545654296875,75.2034184712643
19
+ 1,544,1.52001953125,73.9900785391876
20
+ 2,576,1.463134765625,72.0822596167315
21
+ 2,608,1.376220703125,72.30434210546885
22
+ 2,640,1.373046875,75.35516784317878
23
+ 2,672,1.37646484375,75.00048637711306
24
+ 2,704,1.370849609375,71.54454915080103
25
+ 2,736,1.35595703125,72.10342263835692
26
+ 2,768,1.351318359375,73.61822488939758
27
+ 2,800,1.341552734375,73.69283818050053
28
+ 2,832,1.350341796875,73.35254646396575
29
+ 3,864,1.292724609375,71.73891113636543
30
+ 3,896,1.231201171875,72.50612877979137
31
+ 3,928,1.249267578125,71.73997216828496
32
+ 3,960,1.23486328125,72.24572646319987
33
+ 3,992,1.2470703125,72.21874965045426
34
+ 3,1024,1.260498046875,71.01279770471795
35
+ 3,1056,1.262451171875,71.28239282311822
36
+ 3,1088,1.24609375,72.01843871324893
37
+ 3,1120,1.25048828125,71.49440820181441
38
+ 4,1152,1.151611328125,69.90122275478602
39
+ 4,1184,1.157958984375,71.99470697954118
40
+ 4,1216,1.179931640625,69.07210383354008
41
+ 4,1248,1.173828125,67.60162600897702
42
+ 4,1280,1.152587890625,68.50119537305783
43
+ 4,1312,1.176513671875,69.20852332481955
44
+ 4,1344,1.166015625,69.92232791766861
45
+ 4,1376,1.1796875,71.00892896894848
46
+ 5,1408,1.16357421875,72.45555003088666
47
+ 5,1440,1.0804443359375,69.45015860800906
48
+ 5,1472,1.0814208984375,71.00156597627819
49
+ 5,1504,1.1224365234375,70.58322578457057
50
+ 5,1536,1.097412109375,69.09835049406871
51
+ 5,1568,1.097412109375,70.2241604764949
52
+ 5,1600,1.1026611328125,71.35726715857705
53
+ 5,1632,1.1053466796875,69.87296046317874
54
+ 5,1664,1.11376953125,70.50921749290218
55
+ 6,1696,1.092529296875,70.10898081907449
56
+ 6,1728,1.0654296875,70.09196262215698
57
+ 6,1760,1.068603515625,69.50168151188232
58
+ 6,1792,1.0595703125,68.94335544010495
59
+ 6,1824,1.048583984375,68.59140278421327
60
+ 6,1856,1.0460205078125,69.62795505629208
61
+ 6,1888,1.0633544921875,69.44217502657546
62
+ 6,1920,1.0531005859375,69.53616592289175
63
+ 6,1952,1.0594482421875,69.40128536108074
64
+ 7,1984,1.044189453125,69.35274684809701
65
+ 7,2016,1.013427734375,69.4939563818511
66
+ 7,2048,1.0462646484375,69.50480224133872
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1da96484868a11dbef063c6a1869053c60133a5bef4ce43945094d73f9df6c00
3
+ size 444897069
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
sts-metrics.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "jsick": 82.7495424766893,
3
+ "jsts-val": 80.8645208379796,
4
+ "jsts-train": 77.85762915212484,
5
+ "avg": 80.49056415559791
6
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "clean_up_tokenization_spaces": true,
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": false,
5
+ "do_subword_tokenize": true,
6
+ "do_word_tokenize": true,
7
+ "jumanpp_kwargs": null,
8
+ "mask_token": "[MASK]",
9
+ "mecab_kwargs": {
10
+ "mecab_dic": "unidic_lite"
11
+ },
12
+ "model_max_length": 512,
13
+ "never_split": null,
14
+ "pad_token": "[PAD]",
15
+ "sep_token": "[SEP]",
16
+ "subword_tokenizer_type": "wordpiece",
17
+ "sudachi_kwargs": null,
18
+ "tokenizer_class": "BertJapaneseTokenizer",
19
+ "unk_token": "[UNK]",
20
+ "word_tokenizer_type": "mecab"
21
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff