javiervela commited on
Commit
4182096
1 Parent(s): 4bc9363

First model version OAEI 22 final

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
2_Dense/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"in_features": 768, "out_features": 512, "bias": true, "activation_function": "torch.nn.modules.activation.Tanh"}
2_Dense/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae82d1fe6d2daf671ea46181505c6d0b969a7d0f5947e072488d84d2491dc1e4
3
+ size 1575975
README.md CHANGED
@@ -1,3 +1,88 @@
1
  ---
2
- license: gpl-3.0
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ pipeline_tag: sentence-similarity
3
+ tags:
4
+ - sentence-transformers
5
+ - feature-extraction
6
+ - sentence-similarity
7
+
8
  ---
9
+
10
+ # {MODEL_NAME}
11
+
12
+ This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 512 dimensional dense vector space and can be used for tasks like clustering or semantic search.
13
+
14
+ <!--- Describe your model here -->
15
+
16
+ ## Usage (Sentence-Transformers)
17
+
18
+ Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
19
+
20
+ ```
21
+ pip install -U sentence-transformers
22
+ ```
23
+
24
+ Then you can use the model like this:
25
+
26
+ ```python
27
+ from sentence_transformers import SentenceTransformer
28
+ sentences = ["This is an example sentence", "Each sentence is converted"]
29
+
30
+ model = SentenceTransformer('{MODEL_NAME}')
31
+ embeddings = model.encode(sentences)
32
+ print(embeddings)
33
+ ```
34
+
35
+
36
+
37
+ ## Evaluation Results
38
+
39
+ <!--- Describe how your model was evaluated -->
40
+
41
+ For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
42
+
43
+
44
+ ## Training
45
+ The model was trained with the parameters:
46
+
47
+ **DataLoader**:
48
+
49
+ `torch.utils.data.dataloader.DataLoader` of length 769 with parameters:
50
+ ```
51
+ {'batch_size': 64, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
52
+ ```
53
+
54
+ **Loss**:
55
+
56
+ `sentence_transformers.losses.CosineSimilarityLoss.CosineSimilarityLoss`
57
+
58
+ Parameters of the fit()-Method:
59
+ ```
60
+ {
61
+ "epochs": 3,
62
+ "evaluation_steps": 100,
63
+ "evaluator": "sentence_transformers.evaluation.BinaryClassificationEvaluator.BinaryClassificationEvaluator",
64
+ "max_grad_norm": 1,
65
+ "optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
66
+ "optimizer_params": {
67
+ "lr": 2e-05
68
+ },
69
+ "scheduler": "WarmupLinear",
70
+ "steps_per_epoch": null,
71
+ "warmup_steps": 231,
72
+ "weight_decay": 0.01
73
+ }
74
+ ```
75
+
76
+
77
+ ## Full Model Architecture
78
+ ```
79
+ SentenceTransformer(
80
+ (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: DistilBertModel
81
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
82
+ (2): Dense({'in_features': 768, 'out_features': 512, 'bias': True, 'activation_function': 'torch.nn.modules.activation.Tanh'})
83
+ )
84
+ ```
85
+
86
+ ## Citing & Authors
87
+
88
+ <!--- Describe where people can find more information -->
binary_classification_evaluation_post-test_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cossim_accuracy,cossim_accuracy_threshold,cossim_f1,cossim_precision,cossim_recall,cossim_f1_threshold,cossim_ap,manhattan_accuracy,manhattan_accuracy_threshold,manhattan_f1,manhattan_precision,manhattan_recall,manhattan_f1_threshold,manhattan_ap,euclidean_accuracy,euclidean_accuracy_threshold,euclidean_f1,euclidean_precision,euclidean_recall,euclidean_f1_threshold,euclidean_ap,dot_accuracy,dot_accuracy_threshold,dot_f1,dot_precision,dot_recall,dot_f1_threshold,dot_ap
2
+ -1,-1,0.9880438392560611,0.6683676242828369,0.9881512130900884,0.9850942158057561,0.9912272427129516,0.6648023128509521,0.9906840374622705,0.9847701285761731,12.127042770385742,0.9848891399519842,0.9829919188122533,0.9867936987076691,12.127042770385742,0.9875615035152805,0.9851022441523936,0.7104722261428833,0.985249906050357,0.9812856741835876,0.989246297519102,0.7116281390190125,0.9877012827830944,0.9699672628932011,0.1859782636165619,0.9705977983185471,0.9560761346998536,0.9855673993019527,0.1859782636165619,0.9870584581181674
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/root/.cache/torch/sentence_transformers/sentence-transformers_distiluse-base-multilingual-cased-v2/",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertModel"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "initializer_range": 0.02,
12
+ "max_position_embeddings": 512,
13
+ "model_type": "distilbert",
14
+ "n_heads": 12,
15
+ "n_layers": 6,
16
+ "output_hidden_states": true,
17
+ "output_past": true,
18
+ "pad_token_id": 0,
19
+ "qa_dropout": 0.1,
20
+ "seq_classif_dropout": 0.2,
21
+ "sinusoidal_pos_embds": false,
22
+ "tie_weights_": true,
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.21.2",
25
+ "vocab_size": 119547
26
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.0.0",
4
+ "transformers": "4.7.0",
5
+ "pytorch": "1.9.0+cu102"
6
+ }
7
+ }
eval/binary_classification_evaluation_pre-test_results.csv ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,cossim_accuracy,cossim_accuracy_threshold,cossim_f1,cossim_precision,cossim_recall,cossim_f1_threshold,cossim_ap,manhattan_accuracy,manhattan_accuracy_threshold,manhattan_f1,manhattan_precision,manhattan_recall,manhattan_f1_threshold,manhattan_ap,euclidean_accuracy,euclidean_accuracy_threshold,euclidean_f1,euclidean_precision,euclidean_recall,euclidean_f1_threshold,euclidean_ap,dot_accuracy,dot_accuracy_threshold,dot_f1,dot_precision,dot_recall,dot_f1_threshold,dot_ap
2
+ 0,100,0.9524600275181477,0.4672946333885193,0.9534256763038021,0.9397965728947127,0.967455900386756,0.4663447141647339,0.9844319505886627,0.9412629880912844,15.20600700378418,0.94168079894479,0.9405288416298109,0.9428355815489105,15.227619171142578,0.9818351447490214,0.9416425487498221,0.8699025511741638,0.9426176059618072,0.9309964118134143,0.954532591264975,0.8763056993484497,0.9821682022805556,0.937704606917493,0.3225811719894409,0.9385391369325611,0.929423735084636,0.9478351098952928,0.3163299262523651,0.9709001953055912
3
+ 0,200,0.9680220145181951,0.5949254631996155,0.9683538360409427,0.964008600542208,0.9727384209036883,0.5935291051864624,0.987007440468899,0.9652227546614793,14.084165573120117,0.965781242705756,0.9560073937153419,0.9757570040562211,14.748043060302734,0.9864887112523777,0.9659818759785548,0.8195900917053223,0.9664749614251648,0.9581865381049509,0.9749080275445713,0.8195900917053223,0.9865326997342712,0.9421169995729942,0.31520918011665344,0.9439492786915371,0.9201074787281683,0.969059522686539,0.31520918011665344,0.9721944290820606
4
+ 0,300,0.9743796555487024,0.5207890272140503,0.9747285660801197,0.96721463731773,0.9823601547023866,0.5207890272140503,0.9884642205351162,0.9716278407743038,13.988382339477539,0.972014226881318,0.964521222253181,0.9796245637204037,14.005314826965332,0.9857921718883034,0.9720074014328415,0.7672507166862488,0.972316066066066,0.9673233124824946,0.9773606263560042,0.7672507166862488,0.9859435656417848,0.9504673340608246,0.1996966004371643,0.951889400921659,0.9305342823677809,0.9742477124799547,0.1996966004371643,0.9774554937813437
5
+ 0,400,0.9801205105090858,0.6340268850326538,0.9802851362160637,0.9779384153210665,0.9826431468729365,0.6196614503860474,0.9893568443084639,0.9767993547468805,13.553163528442383,0.9770260747004933,0.9732309996256084,0.9808508631261201,13.646591186523438,0.9863748892107236,0.9769891350761494,0.7490293979644775,0.9771723813997182,0.9731499672560576,0.9812281860201868,0.7656305432319641,0.9865048658228681,0.9560658537742563,0.20362402498722076,0.9575268324006972,0.9318810820462459,0.9846240920667861,0.20362402498722076,0.9798740461917637
6
+ 0,500,0.9829672154481188,0.5959140658378601,0.9831463311581615,0.9785981308411215,0.9877370059428355,0.5943148136138916,0.9894594550181663,0.9780803719694454,13.585075378417969,0.9783261399887409,0.9731217918805413,0.983586454108103,13.585075378417969,0.9861111990375544,0.9780329268871282,0.7329397201538086,0.9782883939038688,0.9726781051846326,0.9839637770021696,0.7661303877830505,0.9862977497705745,0.9557337381980358,0.1837572157382965,0.9569232189851793,0.9371495749683487,0.9775492878030374,0.1834469586610794,0.9817700442867096
7
+ 0,600,0.9834416662712909,0.6344313621520996,0.9836003947182933,0.9799625468164794,0.9872653523252524,0.6123299598693848,0.9900373273106446,0.9797409498505479,13.567154884338379,0.9799181677091661,0.9771149878071657,0.9827374775964531,13.567154884338379,0.9864163885361159,0.9798832850974997,0.7481762170791626,0.9800376647834275,0.97828743302942,0.9817941703612867,0.7525607347488403,0.986708081880904,0.959054893960241,0.2318517565727234,0.9598697977214601,0.9465333822450477,0.9735873974153382,0.2318517565727234,0.9839248581450718
8
+ 0,700,0.9843431228353181,0.6274487972259521,0.9844983089064262,0.9805371011509311,0.9884916517309688,0.6264643669128418,0.9896592266742127,0.9803577359206718,13.767041206359863,0.980530474040632,0.9776798274406827,0.9833977926610697,13.767041206359863,0.986202099183501,0.9803577359206718,0.7824857234954834,0.9805542508219821,0.9765179156141828,0.9846240920667861,0.7895118594169617,0.986489502195945,0.9569673103382834,0.20881310105323792,0.9585674478096021,0.9293179805137289,0.9897179511366853,0.20767581462860107,0.9814717710263878
9
+ 0,-1,0.9853869146462969,0.6371789574623108,0.9855317549793311,0.9815663890708337,0.9895292896896519,0.6356459856033325,0.9900367673475505,0.9834416662712909,14.94007682800293,0.9836526300997704,0.9769259397097134,0.9904725969248184,14.94007682800293,0.9871527526655801,0.983678891682877,0.8339020013809204,0.9838815481210758,0.9774695093566707,0.9903782662013018,0.8339020013809204,0.9874203577288643,0.9664088817194098,0.23532293736934662,0.9672707100591716,0.9484180944610643,0.9868880294311857,0.23318755626678467,0.9845906879218373
10
+ 1,100,0.9852920244816624,0.6412638425827026,0.9854637531651506,0.9797668997668998,0.9912272427129516,0.5880334377288818,0.9898728830071654,0.9828723252834843,13.015702247619629,0.9830253444303381,0.9800300018751172,0.9860390529195359,13.259360313415527,0.9863082138602234,0.9827299900365327,0.7289380431175232,0.9829220230834194,0.9777840007467563,0.9881143288369022,0.7742814421653748,0.9865251165611855,0.9654599800730654,0.21358546614646912,0.9661867162099397,0.9516881690914082,0.9811338552966701,0.2100149691104889,0.984305485984523
11
+ 1,200,0.986952602362765,0.6925540566444397,0.9870691681948558,0.9840615038439903,0.9900952740307518,0.6925540566444397,0.9901630865147804,0.9831569957773877,13.549335479736328,0.9833262881029543,0.9792329279700654,0.9874540137722856,13.549335479736328,0.9872695460985588,0.9832518859420221,0.7404623031616211,0.9833937056028603,0.9808558558558559,0.9859447221960193,0.7407616376876831,0.9874781221802105,0.9649855292498932,0.21055088937282562,0.9659494079159563,0.9441542064492884,0.9887746439015187,0.2036508023738861,0.9843103891953175
12
+ 1,300,0.9875693884328889,0.6685136556625366,0.9876775467971027,0.9848982271831911,0.9904725969248184,0.6611049175262451,0.9901300705322019,0.9829672154481188,13.065546035766602,0.9831352468642834,0.9792251544076361,0.987076690878219,13.065546035766602,0.987279423869607,0.9827299900365327,0.7276773452758789,0.9828915209625869,0.9794847775175644,0.9863220450900858,0.7276773452758789,0.9874692940960793,0.9685913555060018,0.19318749010562897,0.9692350590203551,0.9552074745809288,0.9836807848316197,0.19265085458755493,0.9853284323646103
13
+ 1,400,0.9881387294206956,0.6651833057403564,0.9882507754488203,0.984733539383722,0.9917932270540515,0.6651833057403564,0.9905834158083135,0.9853869146462969,12.821070671081543,0.9855249553529466,0.9820174206237707,0.9890576360720686,13.216412544250488,0.9874928998758354,0.9854818048109314,0.7142951488494873,0.9855986445783133,0.9834695219310604,0.9877370059428355,0.7220486998558044,0.9877643376506413,0.971675285856621,0.21738117933273315,0.972323953456029,0.9559708295350957,0.989246297519102,0.21683421730995178,0.9858400083652858
14
+ 1,500,0.987806613844475,0.6338107585906982,0.9879359714594188,0.9832741543636704,0.9926422035657013,0.6338107585906982,0.9903004532372055,0.9858613654694691,13.584367752075195,0.9860185793375247,0.9808643703911136,0.9912272427129516,13.584367752075195,0.9871356964756788,0.9860037007164207,0.7415610551834106,0.9861482837958399,0.9817688855646971,0.9905669276483351,0.7452684640884399,0.987359905815467,0.970584048963325,0.23631621897220612,0.9712643678160918,0.9547152619589977,0.9883973210074521,0.20529629290103912,0.9851726290537024
15
+ 1,600,0.987806613844475,0.6709846258163452,0.9879189583039534,0.9846326836581709,0.9912272427129516,0.6629726886749268,0.9898221398446679,0.9846752384115386,13.285828590393066,0.9848562989357212,0.9790268456375839,0.9907555890953683,13.317914962768555,0.9868964664702001,0.9848175736584903,0.7318297624588013,0.9849765258215963,0.9804654640620619,0.9895292896896519,0.732924222946167,0.9871191773520593,0.9671680030364853,0.18056347966194153,0.9679777880610828,0.9500408756471977,0.9866050372606358,0.18056347966194153,0.9851050848207736
16
+ 1,700,0.9874744982682545,0.657825231552124,0.9875881523272213,0.9844409035523479,0.9907555890953683,0.657825231552124,0.989827263395934,0.9843431228353181,12.980645179748535,0.9844632768361582,0.9827051414606636,0.9862277143665692,12.980645179748535,0.9870037041565137,0.9843431228353181,0.7193360328674316,0.9844588866911558,0.9829775228063575,0.9859447221960193,0.7255994081497192,0.9871993330475484,0.9685913555060018,0.19866621494293213,0.9691921072226358,0.956461835216313,0.98226582397887,0.19866621494293213,0.9864968446556598
17
+ 1,-1,0.9879963941737439,0.6780102849006653,0.9881025158711497,0.9851837959489872,0.9910385812659183,0.6726783514022827,0.9902680087518279,0.9825402097072639,13.1676025390625,0.982734353007413,0.9775973116774013,0.9879256673898689,13.18469524383545,0.9867682033591311,0.9827299900365327,0.7335050702095032,0.982905982905983,0.9786776395772936,0.9871710216017356,0.7335050702095032,0.9869600460207208,0.9683066850120985,0.17922572791576385,0.9690396737115313,0.9525284738041002,0.9861333836430526,0.16910864412784576,0.9858832192293359
18
+ 2,100,0.9880438392560611,0.6683676242828369,0.9881512130900884,0.9850942158057561,0.9912272427129516,0.6648023128509521,0.9906840374622705,0.9847701285761731,12.127042770385742,0.9848891399519842,0.9829919188122533,0.9867936987076691,12.127042770385742,0.9875615035152805,0.9851022441523936,0.7104722261428833,0.985249906050357,0.9812856741835876,0.989246297519102,0.7116281390190125,0.9877012827830944,0.9699672628932011,0.1859782636165619,0.9705977983185471,0.9560761346998536,0.9855673993019527,0.1859782636165619,0.9870584581181674
19
+ 2,200,0.9883759548322817,0.670170247554779,0.9884841363102234,0.9851039910061833,0.9918875577775682,0.6662254333496094,0.989632976636439,0.9838212269298288,13.01136589050293,0.9839883551673946,0.9796185489902768,0.9883973210074521,13.01136589050293,0.9866200402446548,0.9840110072590976,0.7356672286987305,0.9841761750481289,0.9798055347793567,0.9885859824544855,0.7356672286987305,0.9868157812987344,0.9696351473169806,0.16521108150482178,0.9702795579084239,0.9555474252263788,0.985473068578436,0.1651148796081543,0.9872365723026211
20
+ 2,300,0.9884708449969161,0.6537653207778931,0.9885802904271818,0.9850159205843791,0.9921705499481182,0.6537653207778931,0.9903156502881456,0.9854818048109314,12.828125,0.9856445862263088,0.9804013065795614,0.9909442505424017,13.070582389831543,0.987126699399375,0.9857664753048346,0.7038590908050537,0.9859088774072334,0.9818505005145477,0.9900009433072352,0.7158759832382202,0.9872883174169651,0.9700147079755184,0.17325101792812347,0.9706210487169952,0.956832554302997,0.9848127535138195,0.17325101792812347,0.9871818769345372
21
+ 2,400,0.9883759548322817,0.6587244272232056,0.9884830536360645,0.9851949025487257,0.9917932270540515,0.6587244272232056,0.9901513590838273,0.9853394695639797,12.603439331054688,0.9854677138691623,0.9826486587882198,0.9883029902839354,12.603439331054688,0.986940577950886,0.9854343597286142,0.7158213257789612,0.9855699177438308,0.9821997376803447,0.988963305348552,0.7158213257789612,0.987116007332059,0.9692081415761257,0.1722998321056366,0.9698462110300609,0.9555942135140084,0.9845297613432695,0.17199552059173584,0.9872737234851108
22
+ 2,500,0.9884233999145988,0.6377514600753784,0.9885338345864662,0.984923681992696,0.9921705499481182,0.6377514600753784,0.9902492358447373,0.9859088105517863,12.417805671691895,0.9860347016504444,0.9830301893868367,0.9890576360720686,12.417805671691895,0.9872636194996202,0.986051145798738,0.6742980480194092,0.986156888595913,0.984488107549121,0.9878313366663523,0.6742980480194092,0.9874181760094549,0.9699672628932011,0.18293534219264984,0.9704771232685042,0.9597785977859778,0.98141684746722,0.18293534219264984,0.9866234736426251
23
+ 2,600,0.9886606253261849,0.6834684610366821,0.9887608746766988,0.9858402100525131,0.9916988963305349,0.6834684610366821,0.9901761755433403,0.9858613654694691,12.14111042022705,0.9859949243349939,0.9824857169616934,0.9895292896896519,12.26793098449707,0.987264495624926,0.9860985908810552,0.6728010177612305,0.986215008233357,0.9837619673362118,0.9886803131780021,0.6740717887878418,0.987429528214226,0.9710584997864972,0.16978688538074493,0.9716147045137274,0.958765726880338,0.9848127535138195,0.16978688538074493,0.9868413588173136
24
+ 2,700,0.9887080704085022,0.6865813732147217,0.9888073739653874,0.9859326643533715,0.9916988963305349,0.6865813732147217,0.9903734087661259,0.9860985908810552,11.839252471923828,0.9862124135334808,0.983943661971831,0.9884916517309688,11.839252471923828,0.9873790984196866,0.9861934810456896,0.6626368761062622,0.9863026594492822,0.9843104096204435,0.9883029902839354,0.6656437516212463,0.9875598004998706,0.971248280115766,0.16063669323921204,0.971902818991098,0.9556852375307742,0.9886803131780021,0.1594233512878418,0.9868564922271803
25
+ 2,-1,0.9886606253261849,0.7485941052436829,0.9887608746766988,0.9858402100525131,0.9916988963305349,0.6847038269042969,0.9904193838701356,0.9861460359633724,11.88152027130127,0.9862614096170134,0.9838543133389656,0.9886803131780021,11.88152027130127,0.9874143883381251,0.9863358162926412,0.6642154455184937,0.9864432310299379,0.9844968523912431,0.9883973210074521,0.6652257442474365,0.9875888540798109,0.9715329506096693,0.16328157484531403,0.972162939593579,0.9565415867798777,0.9883029902839354,0.16328157484531403,0.9869493160093921
modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Dense",
18
+ "type": "sentence_transformers.models.Dense"
19
+ }
20
+ ]
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6eafe96420db47b94c174f166b6e68812ac790735bccfdf2f5b3ccf94f1faaf4
3
+ size 538968313
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 128,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_basic_tokenize": true,
4
+ "do_lower_case": false,
5
+ "full_tokenizer_file": null,
6
+ "mask_token": "[MASK]",
7
+ "max_len": 512,
8
+ "name_or_path": "/root/.cache/torch/sentence_transformers/sentence-transformers_distiluse-base-multilingual-cased-v2/",
9
+ "never_split": null,
10
+ "pad_token": "[PAD]",
11
+ "sep_token": "[SEP]",
12
+ "special_tokens_map_file": "/home/reimers/.cache/torch/sentence_transformers/sbert.net_models_distiluse-base-multilingual-cased/0_DistilBERT/special_tokens_map.json",
13
+ "strip_accents": null,
14
+ "tokenize_chinese_chars": true,
15
+ "tokenizer_class": "DistilBertTokenizer",
16
+ "unk_token": "[UNK]"
17
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff