uaritm commited on
Commit
12211cb
1 Parent(s): ffce954

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
README.md CHANGED
@@ -1,3 +1,127 @@
1
  ---
2
- license: mit
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ pipeline_tag: sentence-similarity
3
+ tags:
4
+ - sentence-transformers
5
+ - feature-extraction
6
+ - sentence-similarity
7
+ - transformers
8
+
9
  ---
10
+
11
+ # {MODEL_NAME}
12
+
13
+ This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
14
+
15
+ <!--- Describe your model here -->
16
+
17
+ ## Usage (Sentence-Transformers)
18
+
19
+ Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
20
+
21
+ ```
22
+ pip install -U sentence-transformers
23
+ ```
24
+
25
+ Then you can use the model like this:
26
+
27
+ ```python
28
+ from sentence_transformers import SentenceTransformer
29
+ sentences = ["This is an example sentence", "Each sentence is converted"]
30
+
31
+ model = SentenceTransformer('{MODEL_NAME}')
32
+ embeddings = model.encode(sentences)
33
+ print(embeddings)
34
+ ```
35
+
36
+
37
+
38
+ ## Usage (HuggingFace Transformers)
39
+ Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
40
+
41
+ ```python
42
+ from transformers import AutoTokenizer, AutoModel
43
+ import torch
44
+
45
+
46
+ #Mean Pooling - Take attention mask into account for correct averaging
47
+ def mean_pooling(model_output, attention_mask):
48
+ token_embeddings = model_output[0] #First element of model_output contains all token embeddings
49
+ input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
50
+ return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)
51
+
52
+
53
+ # Sentences we want sentence embeddings for
54
+ sentences = ['This is an example sentence', 'Each sentence is converted']
55
+
56
+ # Load model from HuggingFace Hub
57
+ tokenizer = AutoTokenizer.from_pretrained('{MODEL_NAME}')
58
+ model = AutoModel.from_pretrained('{MODEL_NAME}')
59
+
60
+ # Tokenize sentences
61
+ encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
62
+
63
+ # Compute token embeddings
64
+ with torch.no_grad():
65
+ model_output = model(**encoded_input)
66
+
67
+ # Perform pooling. In this case, mean pooling.
68
+ sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask'])
69
+
70
+ print("Sentence embeddings:")
71
+ print(sentence_embeddings)
72
+ ```
73
+
74
+
75
+
76
+ ## Evaluation Results
77
+
78
+ <!--- Describe how your model was evaluated -->
79
+
80
+ For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
81
+
82
+
83
+ ## Training
84
+ The model was trained with the parameters:
85
+
86
+ **DataLoader**:
87
+
88
+ `torch.utils.data.dataloader.DataLoader` of length 22369 with parameters:
89
+ ```
90
+ {'batch_size': 64, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
91
+ ```
92
+
93
+ **Loss**:
94
+
95
+ `sentence_transformers.losses.MSELoss.MSELoss`
96
+
97
+ Parameters of the fit()-Method:
98
+ ```
99
+ {
100
+ "epochs": 8,
101
+ "evaluation_steps": 1000,
102
+ "evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator",
103
+ "max_grad_norm": 1,
104
+ "optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
105
+ "optimizer_params": {
106
+ "eps": 1e-06,
107
+ "lr": 2e-05
108
+ },
109
+ "scheduler": "WarmupLinear",
110
+ "steps_per_epoch": null,
111
+ "warmup_steps": 1000,
112
+ "weight_decay": 0.01
113
+ }
114
+ ```
115
+
116
+
117
+ ## Full Model Architecture
118
+ ```
119
+ SentenceTransformer(
120
+ (0): Transformer({'max_seq_length': 384, 'do_lower_case': False}) with Transformer model: XLMRobertaModel
121
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
122
+ )
123
+ ```
124
+
125
+ ## Citing & Authors
126
+
127
+ <!--- Describe where people can find more information -->
config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "xlm-roberta-base",
3
+ "architectures": [
4
+ "XLMRobertaModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 514,
17
+ "model_type": "xlm-roberta",
18
+ "num_attention_heads": 12,
19
+ "num_hidden_layers": 12,
20
+ "output_past": true,
21
+ "pad_token_id": 1,
22
+ "position_embedding_type": "absolute",
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.35.2",
25
+ "type_vocab_size": 1,
26
+ "use_cache": true,
27
+ "vocab_size": 250002
28
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.2.2",
4
+ "transformers": "4.35.2",
5
+ "pytorch": "2.1.0+cu121"
6
+ }
7
+ }
eval/mse_evaluation_Tatoeba-eng-rus-dev.tsv.gz_results.csv ADDED
@@ -0,0 +1,185 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,MSE
2
+ 0,1000,0.848949421197176
3
+ 0,2000,0.4606836475431919
4
+ 0,3000,0.8414469659328461
5
+ 0,4000,42.11623966693878
6
+ 0,5000,37.60883808135986
7
+ 0,6000,28.293192386627197
8
+ 0,7000,15.338416397571564
9
+ 0,8000,18.78209561109543
10
+ 0,9000,14.22184556722641
11
+ 0,10000,21.08801007270813
12
+ 0,11000,16.04025810956955
13
+ 0,12000,0.15358192613348365
14
+ 0,13000,0.15338454395532608
15
+ 0,14000,0.13217859668657184
16
+ 0,15000,8.262499421834946
17
+ 0,16000,26.38794183731079
18
+ 0,17000,19.209598004817963
19
+ 0,18000,8.74292254447937
20
+ 0,19000,5.29480054974556
21
+ 0,20000,4.526643082499504
22
+ 0,21000,5.248592421412468
23
+ 0,22000,3.183121234178543
24
+ 0,-1,11.969202011823654
25
+ 1,1000,6.24447837471962
26
+ 1,2000,6.333371251821518
27
+ 1,3000,12.141624838113785
28
+ 1,4000,3.921690955758095
29
+ 1,5000,14.211288094520569
30
+ 1,6000,11.02134883403778
31
+ 1,7000,13.251310586929321
32
+ 1,8000,3.8379158824682236
33
+ 1,9000,3.659619018435478
34
+ 1,10000,3.491947054862976
35
+ 1,11000,0.12413397198542953
36
+ 1,12000,0.12565606739372015
37
+ 1,13000,0.12469175271689892
38
+ 1,14000,0.12597381137311459
39
+ 1,15000,0.1254286034964025
40
+ 1,16000,0.12833582004532218
41
+ 1,17000,0.1254844362847507
42
+ 1,18000,0.12527049984782934
43
+ 1,19000,0.12602504575625062
44
+ 1,20000,0.12823607539758086
45
+ 1,21000,0.12702911626547575
46
+ 1,22000,0.12795787770301104
47
+ 1,-1,0.1278354087844491
48
+ 2,1000,0.13138328213244677
49
+ 2,2000,0.12836138484999537
50
+ 2,3000,0.1274282345548272
51
+ 2,4000,0.1273218891583383
52
+ 2,5000,0.12945140479132533
53
+ 2,6000,0.12839996488764882
54
+ 2,7000,0.12647401308640838
55
+ 2,8000,0.12723514810204506
56
+ 2,9000,0.13070841087028384
57
+ 2,10000,0.12705009430646896
58
+ 2,11000,0.12638139305636287
59
+ 2,12000,0.12608106480911374
60
+ 2,13000,0.12752475449815392
61
+ 2,14000,0.1281437580473721
62
+ 2,15000,0.12951804092153907
63
+ 2,16000,0.12963052140548825
64
+ 2,17000,0.12746391585096717
65
+ 2,18000,0.12532207183539867
66
+ 2,19000,0.1257385010831058
67
+ 2,20000,0.12523052282631397
68
+ 2,21000,0.12620625784620643
69
+ 2,22000,0.12513790279626846
70
+ 2,-1,0.12589959660544991
71
+ 3,1000,0.12622427893802524
72
+ 3,2000,0.12535502901300788
73
+ 3,3000,0.1261907978914678
74
+ 3,4000,0.12519396841526031
75
+ 3,5000,0.12532700784504414
76
+ 3,6000,0.12554720742627978
77
+ 3,7000,0.12509647058323026
78
+ 3,8000,0.12553180567920208
79
+ 3,9000,0.12640993809327483
80
+ 3,10000,0.12675226898863912
81
+ 3,11000,0.12640354689210653
82
+ 3,12000,0.12812383938580751
83
+ 3,13000,0.12684185057878494
84
+ 3,14000,0.12931845849379897
85
+ 3,15000,0.12905217008665204
86
+ 3,16000,0.12640327913686633
87
+ 3,17000,0.12504806509241462
88
+ 3,18000,0.12564064236357808
89
+ 3,19000,0.12659903150051832
90
+ 3,20000,0.12494247639551759
91
+ 3,21000,0.12556747533380985
92
+ 3,22000,0.12491807574406266
93
+ 3,-1,0.12564186472445726
94
+ 4,1000,0.12547316728159785
95
+ 4,2000,0.1254525501281023
96
+ 4,3000,0.12541613541543484
97
+ 4,4000,0.12554129352793097
98
+ 4,5000,0.12682921951636672
99
+ 4,6000,0.1258071744814515
100
+ 4,7000,0.1253918744623661
101
+ 4,8000,0.12559785973280668
102
+ 4,9000,0.12527299113571644
103
+ 4,10000,0.12599037727341056
104
+ 4,11000,0.12512532994151115
105
+ 4,12000,0.125212618149817
106
+ 4,13000,0.12555598514154553
107
+ 4,14000,0.1251509878784418
108
+ 4,15000,0.12564953649416566
109
+ 4,16000,0.12506804196164012
110
+ 4,17000,0.12520997552201152
111
+ 4,18000,0.12657485203817487
112
+ 4,19000,0.12641792418435216
113
+ 4,20000,0.12666251277551055
114
+ 4,21000,0.12724685948342085
115
+ 4,22000,0.12806769227609038
116
+ 4,-1,0.12863221345469356
117
+ 5,1000,0.12642602669075131
118
+ 5,2000,0.12800676049664617
119
+ 5,3000,0.12629072880372405
120
+ 5,4000,0.1290170243009925
121
+ 5,5000,0.12826438760384917
122
+ 5,6000,0.12519805459305644
123
+ 5,7000,0.125317950733006
124
+ 5,8000,0.12615734012797475
125
+ 5,9000,0.12642822694033384
126
+ 5,10000,0.12593038845807314
127
+ 5,11000,0.12542834738269448
128
+ 5,12000,0.12669269926846027
129
+ 5,13000,0.12542062904685736
130
+ 5,14000,0.12645676033571362
131
+ 5,15000,0.12655231403186917
132
+ 5,16000,0.126134033780545
133
+ 5,17000,0.12598291505128145
134
+ 5,18000,0.12577302986755967
135
+ 5,19000,0.125892402138561
136
+ 5,20000,0.12707466958090663
137
+ 5,21000,0.12666366528719664
138
+ 5,22000,0.12758017983287573
139
+ 5,-1,0.12715521734207869
140
+ 6,1000,0.12620933121070266
141
+ 6,2000,0.12544727651402354
142
+ 6,3000,0.12546692742034793
143
+ 6,4000,0.12546839425340295
144
+ 6,5000,0.1252964953891933
145
+ 6,6000,0.12533211847767234
146
+ 6,7000,0.1251009409315884
147
+ 6,8000,0.12540931347757578
148
+ 6,9000,0.12469106586650014
149
+ 6,10000,0.1251082867383957
150
+ 6,11000,0.12571561383083463
151
+ 6,12000,0.1252957619726658
152
+ 6,13000,0.12519304873421788
153
+ 6,14000,0.1253728405572474
154
+ 6,15000,0.12530144304037094
155
+ 6,16000,0.12525240890681744
156
+ 6,17000,0.12537494767457247
157
+ 6,18000,0.12520120944827795
158
+ 6,19000,0.12507732026278973
159
+ 6,20000,0.12561904732137918
160
+ 6,21000,0.12518406147137284
161
+ 6,22000,0.1257007708773017
162
+ 6,-1,0.12541512260213494
163
+ 7,1000,0.12544342316687107
164
+ 7,2000,0.12556433212012053
165
+ 7,3000,0.12590863043442369
166
+ 7,4000,0.12514503905549645
167
+ 7,5000,0.1253378577530384
168
+ 7,6000,0.12566092191264033
169
+ 7,7000,0.1251959940418601
170
+ 7,8000,0.12585095828399062
171
+ 7,9000,0.12555383145809174
172
+ 7,10000,0.12559937313199043
173
+ 7,11000,0.12625348754227161
174
+ 7,12000,0.12589046964421868
175
+ 7,13000,0.125878828112036
176
+ 7,14000,0.12598383473232388
177
+ 7,15000,0.12609409168362617
178
+ 7,16000,0.12586734956130385
179
+ 7,17000,0.12592717539519072
180
+ 7,18000,0.12580857146531343
181
+ 7,19000,0.12572628911584616
182
+ 7,20000,0.12575872242450714
183
+ 7,21000,0.12580197071656585
184
+ 7,22000,0.12579445028677583
185
+ 7,-1,0.12579462490975857
eval/mse_evaluation_Tatoeba-eng-ukr-dev.tsv.gz_results.csv ADDED
@@ -0,0 +1,185 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,MSE
2
+ 0,1000,0.8054990321397781
3
+ 0,2000,0.47793304547667503
4
+ 0,3000,0.6615580059587955
5
+ 0,4000,42.05380976200104
6
+ 0,5000,37.66351044178009
7
+ 0,6000,28.103342652320862
8
+ 0,7000,14.893744885921478
9
+ 0,8000,18.412677943706512
10
+ 0,9000,14.194142818450928
11
+ 0,10000,20.698603987693787
12
+ 0,11000,14.665719866752625
13
+ 0,12000,0.1389687997289002
14
+ 0,13000,0.15049357898533344
15
+ 0,14000,0.13093536254018545
16
+ 0,15000,8.95860269665718
17
+ 0,16000,26.40240788459778
18
+ 0,17000,19.193385541439056
19
+ 0,18000,8.742277324199677
20
+ 0,19000,5.295905098319054
21
+ 0,20000,4.525310918688774
22
+ 0,21000,5.251230299472809
23
+ 0,22000,3.1860072165727615
24
+ 0,-1,11.993832886219025
25
+ 1,1000,6.250041723251343
26
+ 1,2000,6.328821927309036
27
+ 1,3000,12.175609171390533
28
+ 1,4000,3.9220526814460754
29
+ 1,5000,14.225760102272034
30
+ 1,6000,11.053801327943802
31
+ 1,7000,13.287901878356934
32
+ 1,8000,3.8226719945669174
33
+ 1,9000,3.6519233137369156
34
+ 1,10000,3.4523718059062958
35
+ 1,11000,0.12269968865439296
36
+ 1,12000,0.12397165410220623
37
+ 1,13000,0.12290896847844124
38
+ 1,14000,0.12419394915923476
39
+ 1,15000,0.12384355068206787
40
+ 1,16000,0.1259002834558487
41
+ 1,17000,0.12366366572678089
42
+ 1,18000,0.12348245363682508
43
+ 1,19000,0.12393235228955746
44
+ 1,20000,0.12636043829843402
45
+ 1,21000,0.12466326588764787
46
+ 1,22000,0.1255976501852274
47
+ 1,-1,0.1256137154996395
48
+ 2,1000,0.1287964405491948
49
+ 2,2000,0.12630567653104663
50
+ 2,3000,0.1253655762411654
51
+ 2,4000,0.12530474923551083
52
+ 2,5000,0.12748391600325704
53
+ 2,6000,0.1263804384507239
54
+ 2,7000,0.12450109934434295
55
+ 2,8000,0.12522073229774833
56
+ 2,9000,0.1290328335016966
57
+ 2,10000,0.1253199647180736
58
+ 2,11000,0.12431163340806961
59
+ 2,12000,0.12405728921294212
60
+ 2,13000,0.12577766319736838
61
+ 2,14000,0.12671578442677855
62
+ 2,15000,0.12822209391742945
63
+ 2,16000,0.12833329383283854
64
+ 2,17000,0.12560287723317742
65
+ 2,18000,0.12336722575128078
66
+ 2,19000,0.1235613483004272
67
+ 2,20000,0.12324543204158545
68
+ 2,21000,0.12404901208356023
69
+ 2,22000,0.12313963379710913
70
+ 2,-1,0.1236567273736
71
+ 3,1000,0.12402331922203302
72
+ 3,2000,0.12336981017142534
73
+ 3,3000,0.12433367082849145
74
+ 3,4000,0.12314142659306526
75
+ 3,5000,0.12318437220528722
76
+ 3,6000,0.12359702959656715
77
+ 3,7000,0.1231974340043962
78
+ 3,8000,0.1236186595633626
79
+ 3,9000,0.12440326390787959
80
+ 3,10000,0.12481571175158024
81
+ 3,11000,0.12431573122739792
82
+ 3,12000,0.12561732437461615
83
+ 3,13000,0.1248013461008668
84
+ 3,14000,0.12722813989967108
85
+ 3,15000,0.1275797374546528
86
+ 3,16000,0.12423582375049591
87
+ 3,17000,0.12318630469962955
88
+ 3,18000,0.12369363103061914
89
+ 3,19000,0.12428282061591744
90
+ 3,20000,0.12305304408073425
91
+ 3,21000,0.12351544573903084
92
+ 3,22000,0.12311561731621623
93
+ 3,-1,0.12360598193481565
94
+ 4,1000,0.12369530741125345
95
+ 4,2000,0.12366650626063347
96
+ 4,3000,0.12371109332889318
97
+ 4,4000,0.12373398058116436
98
+ 4,5000,0.12460004072636366
99
+ 4,6000,0.12386356247588992
100
+ 4,7000,0.12358080130070448
101
+ 4,8000,0.12383370194584131
102
+ 4,9000,0.12345609720796347
103
+ 4,10000,0.12429772177711129
104
+ 4,11000,0.12324347626417875
105
+ 4,12000,0.1233651302754879
106
+ 4,13000,0.12377903331071138
107
+ 4,14000,0.12338090455159545
108
+ 4,15000,0.12387686874717474
109
+ 4,16000,0.123142811935395
110
+ 4,17000,0.12337157968431711
111
+ 4,18000,0.12481666635721922
112
+ 4,19000,0.12445036554709077
113
+ 4,20000,0.12473324313759804
114
+ 4,21000,0.12555629946291447
115
+ 4,22000,0.12646267423406243
116
+ 4,-1,0.1270242384634912
117
+ 5,1000,0.12452059891074896
118
+ 5,2000,0.12606095988303423
119
+ 5,3000,0.12432234361767769
120
+ 5,4000,0.12658301275223494
121
+ 5,5000,0.12622650247067213
122
+ 5,6000,0.12355041690170765
123
+ 5,7000,0.12346578296273947
124
+ 5,8000,0.12455566320568323
125
+ 5,9000,0.12474027462303638
126
+ 5,10000,0.12430166825652122
127
+ 5,11000,0.1238511991687119
128
+ 5,12000,0.12524921912699938
129
+ 5,13000,0.12373232748359442
130
+ 5,14000,0.12470505898818374
131
+ 5,15000,0.12461120495572686
132
+ 5,16000,0.12418084079399705
133
+ 5,17000,0.12404794106259942
134
+ 5,18000,0.12389132753014565
135
+ 5,19000,0.12383211869746447
136
+ 5,20000,0.12506392085924745
137
+ 5,21000,0.12490013614296913
138
+ 5,22000,0.12559752212837338
139
+ 5,-1,0.1251919544301927
140
+ 6,1000,0.12436301913112402
141
+ 6,2000,0.12355090584605932
142
+ 6,3000,0.12357770465314388
143
+ 6,4000,0.12361875269562006
144
+ 6,5000,0.12351239565759897
145
+ 6,6000,0.12347641168162227
146
+ 6,7000,0.12330394238233566
147
+ 6,8000,0.12360374676063657
148
+ 6,9000,0.12283488176763058
149
+ 6,10000,0.12329661985859275
150
+ 6,11000,0.12394378427416086
151
+ 6,12000,0.12342417612671852
152
+ 6,13000,0.12325469870120287
153
+ 6,14000,0.12352274497970939
154
+ 6,15000,0.12338808737695217
155
+ 6,16000,0.123451033141464
156
+ 6,17000,0.12355117360129952
157
+ 6,18000,0.12339792447164655
158
+ 6,19000,0.12331136967986822
159
+ 6,20000,0.12373700737953186
160
+ 6,21000,0.12345025315880775
161
+ 6,22000,0.12411287752911448
162
+ 6,-1,0.12393624056130648
163
+ 7,1000,0.12376323575153947
164
+ 7,2000,0.12389406329020858
165
+ 7,3000,0.1241093035787344
166
+ 7,4000,0.12334435014054179
167
+ 7,5000,0.12357927625998855
168
+ 7,6000,0.12391237542033195
169
+ 7,7000,0.12356614461168647
170
+ 7,8000,0.12426013126969337
171
+ 7,9000,0.12394718360155821
172
+ 7,10000,0.12392947683110833
173
+ 7,11000,0.12458892306312919
174
+ 7,12000,0.12418971164152026
175
+ 7,13000,0.12411647476255894
176
+ 7,14000,0.12425838503986597
177
+ 7,15000,0.12436830438673496
178
+ 7,16000,0.12419078266248107
179
+ 7,17000,0.12426362372934818
180
+ 7,18000,0.12410922208800912
181
+ 7,19000,0.12405223678797483
182
+ 7,20000,0.12411032803356647
183
+ 7,21000,0.12414938537403941
184
+ 7,22000,0.12413804652169347
185
+ 7,-1,0.12413694057613611
eval/translation_evaluation_Tatoeba-eng-rus-dev.tsv.gz_results.csv ADDED
@@ -0,0 +1,185 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,src2trg,trg2src
2
+ 0,1000,0.00195,0.00335
3
+ 0,2000,0.00135,0.0013
4
+ 0,3000,0.00025,0.00055
5
+ 0,4000,0.00025,0.00025
6
+ 0,5000,5e-05,0.00025
7
+ 0,6000,0.0,0.0001
8
+ 0,7000,0.00015,0.00015
9
+ 0,8000,0.0002,0.00025
10
+ 0,9000,0.0001,0.0002
11
+ 0,10000,0.0002,0.00045
12
+ 0,11000,0.00035,0.0004
13
+ 0,12000,0.00015,0.00025
14
+ 0,13000,0.0001,0.0002
15
+ 0,14000,0.0003,0.0001
16
+ 0,15000,0.00015,0.00025
17
+ 0,16000,5e-05,0.0001
18
+ 0,17000,5e-05,5e-05
19
+ 0,18000,0.00015,5e-05
20
+ 0,19000,5e-05,5e-05
21
+ 0,20000,0.0,5e-05
22
+ 0,21000,0.0001,0.00015
23
+ 0,22000,5e-05,0.00025
24
+ 0,-1,0.0001,0.0002
25
+ 1,1000,5e-05,0.0002
26
+ 1,2000,0.00015,0.0001
27
+ 1,3000,0.0,0.0
28
+ 1,4000,0.0001,0.00015
29
+ 1,5000,5e-05,5e-05
30
+ 1,6000,0.0001,0.0002
31
+ 1,7000,0.0,0.0002
32
+ 1,8000,5e-05,0.0003
33
+ 1,9000,5e-05,5e-05
34
+ 1,10000,0.0002,0.0002
35
+ 1,11000,0.0,0.00015
36
+ 1,12000,0.0003,0.0004
37
+ 1,13000,0.0001,0.00025
38
+ 1,14000,0.0002,0.0004
39
+ 1,15000,0.0002,5e-05
40
+ 1,16000,0.0001,0.0002
41
+ 1,17000,0.00025,0.0002
42
+ 1,18000,0.0003,0.00035
43
+ 1,19000,5e-05,0.0001
44
+ 1,20000,0.00025,0.0002
45
+ 1,21000,0.0003,0.0002
46
+ 1,22000,0.00015,0.00025
47
+ 1,-1,0.0001,0.0002
48
+ 2,1000,0.0002,0.0004
49
+ 2,2000,0.00065,0.00045
50
+ 2,3000,0.00025,0.00055
51
+ 2,4000,0.0002,0.0002
52
+ 2,5000,0.00025,0.00035
53
+ 2,6000,0.00025,0.00035
54
+ 2,7000,0.00035,0.00035
55
+ 2,8000,0.0003,0.00045
56
+ 2,9000,0.00035,0.00065
57
+ 2,10000,0.0003,0.00025
58
+ 2,11000,0.0003,0.00035
59
+ 2,12000,0.0003,0.0005
60
+ 2,13000,5e-05,0.00025
61
+ 2,14000,0.00025,0.00015
62
+ 2,15000,0.0003,0.00035
63
+ 2,16000,0.00015,0.00035
64
+ 2,17000,5e-05,0.00015
65
+ 2,18000,0.00015,0.0003
66
+ 2,19000,0.0001,0.0001
67
+ 2,20000,0.00015,0.0002
68
+ 2,21000,0.0001,0.0
69
+ 2,22000,5e-05,0.00025
70
+ 2,-1,5e-05,0.00015
71
+ 3,1000,0.0001,0.00025
72
+ 3,2000,5e-05,0.0
73
+ 3,3000,0.00015,0.0002
74
+ 3,4000,0.0001,5e-05
75
+ 3,5000,0.00015,0.0001
76
+ 3,6000,0.00015,0.0003
77
+ 3,7000,0.00015,0.00045
78
+ 3,8000,0.00025,0.00025
79
+ 3,9000,0.0003,0.0003
80
+ 3,10000,0.0003,0.0003
81
+ 3,11000,0.00045,0.0004
82
+ 3,12000,0.0001,0.0005
83
+ 3,13000,0.0004,0.0004
84
+ 3,14000,0.0003,0.0002
85
+ 3,15000,0.0002,0.0003
86
+ 3,16000,0.00035,0.0005
87
+ 3,17000,0.00035,0.0003
88
+ 3,18000,0.00015,0.00045
89
+ 3,19000,0.0003,0.00055
90
+ 3,20000,0.00025,0.0004
91
+ 3,21000,0.00035,0.00045
92
+ 3,22000,0.0002,0.00025
93
+ 3,-1,0.0002,0.00025
94
+ 4,1000,0.0002,0.00035
95
+ 4,2000,0.0003,0.0003
96
+ 4,3000,0.00025,0.0005
97
+ 4,4000,0.0003,0.00025
98
+ 4,5000,0.0002,0.00035
99
+ 4,6000,0.00035,0.00035
100
+ 4,7000,0.00015,0.0003
101
+ 4,8000,0.00025,0.00035
102
+ 4,9000,0.00025,0.0005
103
+ 4,10000,0.00015,0.0003
104
+ 4,11000,0.0003,0.00055
105
+ 4,12000,0.00025,0.00025
106
+ 4,13000,0.0003,0.0004
107
+ 4,14000,0.00035,0.00035
108
+ 4,15000,0.00015,0.0004
109
+ 4,16000,0.00045,0.00035
110
+ 4,17000,0.00015,0.0003
111
+ 4,18000,0.00015,0.00015
112
+ 4,19000,0.00035,0.0002
113
+ 4,20000,0.0001,0.00015
114
+ 4,21000,0.00025,0.00025
115
+ 4,22000,0.00025,0.00035
116
+ 4,-1,0.00015,0.00015
117
+ 5,1000,0.0003,0.0002
118
+ 5,2000,0.00025,0.00025
119
+ 5,3000,0.00045,0.0004
120
+ 5,4000,0.0001,0.0004
121
+ 5,5000,0.00025,0.00015
122
+ 5,6000,0.00015,0.00045
123
+ 5,7000,0.00035,0.00035
124
+ 5,8000,0.0004,0.00025
125
+ 5,9000,0.00025,0.0003
126
+ 5,10000,0.0002,0.0004
127
+ 5,11000,0.0003,0.00035
128
+ 5,12000,0.00025,0.00045
129
+ 5,13000,0.00035,0.0004
130
+ 5,14000,0.00015,0.0002
131
+ 5,15000,0.0002,0.0002
132
+ 5,16000,0.0001,0.0002
133
+ 5,17000,0.00015,0.0003
134
+ 5,18000,0.0001,0.00015
135
+ 5,19000,0.00025,0.0003
136
+ 5,20000,0.0002,0.0005
137
+ 5,21000,0.0002,0.00035
138
+ 5,22000,5e-05,0.00015
139
+ 5,-1,0.0001,0.0003
140
+ 6,1000,0.0001,0.00025
141
+ 6,2000,0.0003,0.0002
142
+ 6,3000,0.00025,0.0001
143
+ 6,4000,5e-05,0.0002
144
+ 6,5000,0.0001,0.0002
145
+ 6,6000,0.0004,0.00025
146
+ 6,7000,0.0001,0.0003
147
+ 6,8000,0.0002,0.00025
148
+ 6,9000,0.00015,0.0002
149
+ 6,10000,0.0001,0.0002
150
+ 6,11000,0.0003,0.0003
151
+ 6,12000,0.0002,0.00025
152
+ 6,13000,0.0001,0.0002
153
+ 6,14000,0.0003,0.0004
154
+ 6,15000,0.00015,0.00025
155
+ 6,16000,0.0001,0.00015
156
+ 6,17000,0.00015,0.0
157
+ 6,18000,5e-05,0.00025
158
+ 6,19000,0.0001,0.0003
159
+ 6,20000,5e-05,0.00035
160
+ 6,21000,0.0003,0.0005
161
+ 6,22000,0.0003,0.00035
162
+ 6,-1,0.00035,0.00045
163
+ 7,1000,0.00015,0.0002
164
+ 7,2000,0.0002,0.00025
165
+ 7,3000,0.0003,0.00025
166
+ 7,4000,0.00025,0.00015
167
+ 7,5000,0.0003,0.0003
168
+ 7,6000,0.00015,0.0002
169
+ 7,7000,0.0003,0.00025
170
+ 7,8000,0.0003,0.00025
171
+ 7,9000,5e-05,0.00015
172
+ 7,10000,0.0001,0.00025
173
+ 7,11000,0.0,0.00035
174
+ 7,12000,5e-05,0.0003
175
+ 7,13000,5e-05,0.0003
176
+ 7,14000,0.00015,0.0003
177
+ 7,15000,5e-05,0.0003
178
+ 7,16000,5e-05,0.00025
179
+ 7,17000,0.0003,0.00025
180
+ 7,18000,0.0001,0.00025
181
+ 7,19000,0.00015,0.00055
182
+ 7,20000,0.0001,0.0003
183
+ 7,21000,0.0001,0.0002
184
+ 7,22000,0.00015,0.0002
185
+ 7,-1,0.0001,0.0003
eval/translation_evaluation_Tatoeba-eng-ukr-dev.tsv.gz_results.csv ADDED
@@ -0,0 +1,185 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,src2trg,trg2src
2
+ 0,1000,0.0012,0.00155
3
+ 0,2000,0.00045,0.00085
4
+ 0,3000,0.00035,0.0005
5
+ 0,4000,0.00015,0.00035
6
+ 0,5000,0.00015,0.0003
7
+ 0,6000,0.0,0.0001
8
+ 0,7000,0.00035,0.0002
9
+ 0,8000,5e-05,0.00025
10
+ 0,9000,0.00015,0.0002
11
+ 0,10000,0.00015,0.0003
12
+ 0,11000,0.00025,0.0003
13
+ 0,12000,0.0003,0.00035
14
+ 0,13000,0.00015,0.0
15
+ 0,14000,5e-05,0.00025
16
+ 0,15000,0.00015,0.00015
17
+ 0,16000,0.0001,0.0
18
+ 0,17000,0.0,0.0
19
+ 0,18000,0.0001,0.0001
20
+ 0,19000,5e-05,0.0001
21
+ 0,20000,0.0,0.0
22
+ 0,21000,5e-05,5e-05
23
+ 0,22000,0.0002,0.00025
24
+ 0,-1,5e-05,0.0001
25
+ 1,1000,0.0001,5e-05
26
+ 1,2000,5e-05,5e-05
27
+ 1,3000,5e-05,0.0002
28
+ 1,4000,5e-05,0.0001
29
+ 1,5000,0.00015,0.0
30
+ 1,6000,5e-05,0.00015
31
+ 1,7000,5e-05,5e-05
32
+ 1,8000,0.00025,0.00025
33
+ 1,9000,5e-05,5e-05
34
+ 1,10000,0.00015,0.0001
35
+ 1,11000,5e-05,0.00025
36
+ 1,12000,0.0002,0.00025
37
+ 1,13000,0.0003,0.00035
38
+ 1,14000,0.00015,0.0002
39
+ 1,15000,0.0004,0.00025
40
+ 1,16000,0.00035,0.00035
41
+ 1,17000,0.00015,0.00035
42
+ 1,18000,0.00015,0.00015
43
+ 1,19000,0.0001,5e-05
44
+ 1,20000,0.0001,0.0001
45
+ 1,21000,0.00015,0.0003
46
+ 1,22000,0.0002,0.00055
47
+ 1,-1,0.0001,0.0001
48
+ 2,1000,0.00015,0.00015
49
+ 2,2000,0.0001,0.0002
50
+ 2,3000,0.00035,0.00025
51
+ 2,4000,0.00025,0.00015
52
+ 2,5000,0.00025,0.00035
53
+ 2,6000,0.00025,0.00025
54
+ 2,7000,0.0002,0.00025
55
+ 2,8000,0.00025,0.0003
56
+ 2,9000,0.0004,0.00035
57
+ 2,10000,0.0003,0.00015
58
+ 2,11000,0.00025,0.00015
59
+ 2,12000,0.00025,0.0004
60
+ 2,13000,0.00015,0.0003
61
+ 2,14000,0.00025,0.0003
62
+ 2,15000,0.0002,0.00045
63
+ 2,16000,0.0003,0.00055
64
+ 2,17000,0.0001,0.0002
65
+ 2,18000,0.0002,0.00015
66
+ 2,19000,0.0002,0.0001
67
+ 2,20000,0.00035,0.0001
68
+ 2,21000,0.0001,0.0001
69
+ 2,22000,0.0002,0.0003
70
+ 2,-1,0.0001,0.00015
71
+ 3,1000,0.0003,0.0002
72
+ 3,2000,5e-05,0.00015
73
+ 3,3000,0.00015,0.0003
74
+ 3,4000,0.0001,0.0002
75
+ 3,5000,0.00035,0.00025
76
+ 3,6000,5e-05,0.0002
77
+ 3,7000,0.0001,0.00025
78
+ 3,8000,0.00025,0.0002
79
+ 3,9000,0.00025,0.00035
80
+ 3,10000,0.0003,0.00035
81
+ 3,11000,5e-05,0.0002
82
+ 3,12000,0.00045,0.0003
83
+ 3,13000,0.00025,0.00035
84
+ 3,14000,0.0001,0.00015
85
+ 3,15000,0.0003,0.0002
86
+ 3,16000,0.0003,0.00025
87
+ 3,17000,5e-05,5e-05
88
+ 3,18000,0.0002,0.0002
89
+ 3,19000,0.00015,0.00035
90
+ 3,20000,0.0003,0.00025
91
+ 3,21000,0.00035,0.00025
92
+ 3,22000,0.0003,0.00035
93
+ 3,-1,0.0001,5e-05
94
+ 4,1000,0.0002,0.0002
95
+ 4,2000,0.00025,0.00015
96
+ 4,3000,0.0004,0.00055
97
+ 4,4000,0.0002,0.0005
98
+ 4,5000,0.00025,0.00035
99
+ 4,6000,0.0002,0.00025
100
+ 4,7000,0.0003,0.0004
101
+ 4,8000,0.0003,0.00045
102
+ 4,9000,0.00025,0.0005
103
+ 4,10000,0.00025,0.00025
104
+ 4,11000,0.00015,0.00055
105
+ 4,12000,0.00025,0.0004
106
+ 4,13000,0.0003,0.00025
107
+ 4,14000,0.0002,0.00025
108
+ 4,15000,0.0002,0.00035
109
+ 4,16000,0.00015,0.00015
110
+ 4,17000,0.00035,0.0003
111
+ 4,18000,0.0003,0.0003
112
+ 4,19000,0.00025,0.00015
113
+ 4,20000,0.00015,0.0001
114
+ 4,21000,0.00015,0.0003
115
+ 4,22000,0.0002,0.00015
116
+ 4,-1,0.0002,0.00015
117
+ 5,1000,0.0001,0.00045
118
+ 5,2000,0.0002,0.00035
119
+ 5,3000,0.00025,0.0003
120
+ 5,4000,0.0003,0.0003
121
+ 5,5000,0.0002,0.00015
122
+ 5,6000,0.0003,0.0002
123
+ 5,7000,5e-05,0.00045
124
+ 5,8000,0.0002,0.0003
125
+ 5,9000,0.00025,0.0001
126
+ 5,10000,0.0001,0.0001
127
+ 5,11000,0.0003,0.0003
128
+ 5,12000,0.00015,0.00025
129
+ 5,13000,0.0003,0.00035
130
+ 5,14000,5e-05,0.0002
131
+ 5,15000,0.0001,0.0003
132
+ 5,16000,0.0002,0.0004
133
+ 5,17000,0.0002,0.0002
134
+ 5,18000,0.00035,0.00025
135
+ 5,19000,0.00015,0.0004
136
+ 5,20000,0.00025,0.0004
137
+ 5,21000,0.0003,0.00025
138
+ 5,22000,0.0002,0.0004
139
+ 5,-1,0.0001,0.00035
140
+ 6,1000,0.0001,0.00025
141
+ 6,2000,0.0001,0.0002
142
+ 6,3000,0.00015,0.00015
143
+ 6,4000,0.0003,0.0003
144
+ 6,5000,5e-05,0.00025
145
+ 6,6000,0.00025,0.0003
146
+ 6,7000,5e-05,0.00015
147
+ 6,8000,0.0002,0.0004
148
+ 6,9000,0.0001,0.0002
149
+ 6,10000,0.0002,0.0003
150
+ 6,11000,0.0002,0.0002
151
+ 6,12000,0.00025,0.00015
152
+ 6,13000,0.0,0.00015
153
+ 6,14000,0.0001,0.00035
154
+ 6,15000,0.0003,0.00035
155
+ 6,16000,0.0004,0.00025
156
+ 6,17000,0.00015,0.0002
157
+ 6,18000,0.0002,0.00025
158
+ 6,19000,0.00015,0.0002
159
+ 6,20000,0.00015,5e-05
160
+ 6,21000,0.00015,0.00045
161
+ 6,22000,0.00015,5e-05
162
+ 6,-1,0.00015,0.0001
163
+ 7,1000,0.0001,0.0003
164
+ 7,2000,5e-05,0.0004
165
+ 7,3000,0.00015,0.00025
166
+ 7,4000,0.00015,0.00025
167
+ 7,5000,0.0001,0.0003
168
+ 7,6000,0.0002,0.00035
169
+ 7,7000,0.00015,0.00015
170
+ 7,8000,0.0002,0.0002
171
+ 7,9000,5e-05,0.00025
172
+ 7,10000,0.00015,0.00025
173
+ 7,11000,0.0001,0.00025
174
+ 7,12000,0.00025,0.00025
175
+ 7,13000,0.0002,0.0003
176
+ 7,14000,0.0002,0.0002
177
+ 7,15000,0.0001,0.00035
178
+ 7,16000,0.0001,0.00035
179
+ 7,17000,0.0003,0.0002
180
+ 7,18000,0.00025,0.0002
181
+ 7,19000,0.0002,0.00035
182
+ 7,20000,0.00025,0.00035
183
+ 7,21000,0.00025,0.00015
184
+ 7,22000,0.00015,0.0003
185
+ 7,-1,0.0001,0.00025
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cccf452832436f9f6f9292aff96a079f188f30429764e987c60ce947db0dfe09
3
+ size 1112197096
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 384,
3
+ "do_lower_case": false
4
+ }
sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
3
+ size 5069051
special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "content": "<mask>",
7
+ "lstrip": true,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
15
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ab208220fd15ae86d71654fbffe08ba919e926330fa27dc19fe32874e3e0492
3
+ size 17083009
tokenizer_config.json ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "250001": {
36
+ "content": "<mask>",
37
+ "lstrip": true,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "<s>",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "<s>",
47
+ "eos_token": "</s>",
48
+ "mask_token": "<mask>",
49
+ "model_max_length": 512,
50
+ "pad_token": "<pad>",
51
+ "sep_token": "</s>",
52
+ "tokenizer_class": "XLMRobertaTokenizer",
53
+ "unk_token": "<unk>"
54
+ }