ramamimu commited on
Commit
d20f3d5
1 Parent(s): 14b174f

initial commit

Browse files
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ unigram.json filter=lfs diff=lfs merge=lfs -text
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 384,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md CHANGED
@@ -1,3 +1,391 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: []
3
+ library_name: sentence-transformers
4
+ tags:
5
+ - sentence-transformers
6
+ - sentence-similarity
7
+ - feature-extraction
8
+ - generated_from_trainer
9
+ - dataset_size:704
10
+ - loss:MultipleNegativesRankingLoss
11
+ base_model: sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2
12
+ datasets: []
13
+ widget:
14
+ - source_sentence: Apa statistik peserta MSIB Batch 4 dan 5?
15
+ sentences:
16
+ - Ijazah dicetak berdasarkan data preview yang dipermanen oleh calon wisudawan.
17
+ Ijazah dicetak hanya sekali saja, bila ada kekeliruan pengisian data pada point
18
+ 3 maka akan diterbitkan surat keterangan
19
+ - 'MSIB Batch 4: Total Mahasiswa diterima: 664 mahasiswa; Program Magang: 228 mahasiswa;
20
+ Program Studi Independen: 436 mahasiswa; Jumlah Mitra Industri: 108 Mitra
21
+
22
+ MSIB Batch 5: Total Mahasiswa diterima: 488 mahasiswa; Program Magang: 257 mahasiswa;
23
+ Program Studi Independen: 231 mahasiswa; Jumlah Mitra Industri: 121 Mitra'
24
+ - Bentuk kegiatan studi/ proyek independen bisa berupa lomba - lomba kemahasiswaan
25
+ atau proyek - proyek untuk memecahkan persoalan di ITS, di masyarakat atau industri.
26
+ - source_sentence: Apa deskripsi penelitian P12 tentang Numerical modeling and Experiments
27
+ of an atmospheric pressure plasma jet operated in air with shielding gas?
28
+ sentences:
29
+ - 'Deskripsi penelitian: Developing a two-dimensional axisymmetric plasma fluid
30
+ model integrated with a gas flow model to predict the dynamic behavior of a helium
31
+ atmospheric pressure plasma jet.'
32
+ - 'Sekilas Kerja Praktik
33
+
34
+ Jumlah kredit Mata Kuliah KP: 2 SKS'
35
+ - Nama pengarang, tahun publikasi, dan judul artikel/paper sama dengan penulisan
36
+ artikel dari jurnal. Judul buku dicetak miring/italic. Nomor volume dari buku
37
+ (jika ada). Edisi penerbitan. Nama editor didahului dengan ed. atau eds. bila
38
+ lebih dari satu editor. Nama penyelenggara seminar/conference. Kota tempat penyelenggaraan.
39
+ Nomor halaman dari artikel/paper tersebut di dalam prosiding.
40
+ - source_sentence: Bagaimana format penulisan referensi proyek mahasiswa?
41
+ sentences:
42
+ - Gedung KPA, Plaza dr.Angka, Lantai 1, Kampus ITS Sukolilo Surabaya.
43
+ - Nama pengarang dan tahun publikasi sama dengan penulisan artikel dari jurnal.
44
+ Judul proyek dicetak miring/italic. Jenis proyek. Nama perguruan tinggi. Kota
45
+ tempat penyelenggaraan.
46
+ - Nama pengarang dan tahun publikasi sama dengan penulisan artikel dari jurnal.
47
+ Judul standar teknis dicetak miring/italic. Nama penerbit. Kota tempat diterbitkan.
48
+ - source_sentence: MyITS saya bermasalah, bisa lapor kemana?
49
+ sentences:
50
+ - Mahasiswa yang ingin mendaftar sidang proposal tesis, harus melengkapi berkas
51
+ persyaratan pendaftaransidang proposal tesis, meliputi draft proposal tesis dan
52
+ lembar persetujuan pembimbing.
53
+ - 'B. Bagi Dosen Wali
54
+
55
+ 2. Jika dinilai sudah sesuai, pengajuan perencanaan Kegiatan SKEM/ Program MB-KM
56
+ oleh mahasiswa, dosen wali dapat melakukan persetujuan FRS.'
57
+ - Silakan ajukan tiket ke DPTSI di https://servicedesk.its.ac.id/.
58
+ - source_sentence: Apa saja panduan umum pelaksanaan Program Magang di ITS?
59
+ sentences:
60
+ - 'Mahasiswa dalam melaksanakan magang harus memenuhi ketentuan berikut: 1. Pelaksanaan
61
+ Program Magang memiliki durasi minimal 1 bulan dan maksimal 6 bulan. 2. Selama
62
+ Program Magang berlangsung, mahasiswa tidak harus mengajukan cuti. 3. Mahasiswa
63
+ secara penuh waktu bekerja di lapangan sesuai kesepakatan. 4. Mahasiswa bisa mendapatkan
64
+ izin untuk kegiatan akademik tertentu dengan kesepakatan pihak Mitra Magang. 5.
65
+ Mahasiswa dapat mengajukan konversi mata kuliah dengan CPMK yang selaras. 6. Diperlukan
66
+ pembimbing internal dari Dosen Departemen dan pembimbing lapangan dari Mitra Magang.
67
+ 7. Sebelum Program Magang, wajib menandatangani perjanjian kerjasama dan nota
68
+ kesepahaman.'
69
+ - Pelaksanaan transfer kegiatan MB - KM menjadi sks mata kuliah, program studi atau
70
+ direktorat membentuk tim pelaksana transfer kredit.
71
+ - '1. Mahasiswa mengurus surat rekomendasi departemen untuk pengajuan magang ke
72
+ Mitra
73
+
74
+ 2. Mahasiswa mengajukan permohonan magang ke Mitra
75
+
76
+ 3. Mitra melakukan seleksi magang
77
+
78
+ 4. Mahasiswa menerima hasil seleksi magang dari Mitra
79
+
80
+ 5. Apabila tidak diterima, maka mahasiswa harus mengulang sejak langkah awal
81
+
82
+ 6. Apabila diterima, Mahasiswa melaporkan ke Departemen
83
+
84
+ 7. Mahasiswa/Departemen melakukan koordinasi dengan PK2 untuk pengurusan PKS dengan
85
+ menyertakan Proposal Magang, Surat Rekomendasi Departemen dan Surat Penerimaan
86
+ Magang
87
+
88
+ 8. Proses pengurusan PKS (Dapat dilakukan bersamaan dengan pelaksanaan Magang)'
89
+ pipeline_tag: sentence-similarity
90
+ ---
91
+
92
+ # SentenceTransformer based on sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2
93
+
94
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
95
+
96
+ ## Model Details
97
+
98
+ ### Model Description
99
+ - **Model Type:** Sentence Transformer
100
+ - **Base model:** [sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2) <!-- at revision bf3bf13ab40c3157080a7ab344c831b9ad18b5eb -->
101
+ - **Maximum Sequence Length:** 128 tokens
102
+ - **Output Dimensionality:** 384 tokens
103
+ - **Similarity Function:** Cosine Similarity
104
+ <!-- - **Training Dataset:** Unknown -->
105
+ <!-- - **Language:** Unknown -->
106
+ <!-- - **License:** Unknown -->
107
+
108
+ ### Model Sources
109
+
110
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
111
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
112
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
113
+
114
+ ### Full Model Architecture
115
+
116
+ ```
117
+ SentenceTransformer(
118
+ (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: BertModel
119
+ (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
120
+ )
121
+ ```
122
+
123
+ ## Usage
124
+
125
+ ### Direct Usage (Sentence Transformers)
126
+
127
+ First install the Sentence Transformers library:
128
+
129
+ ```bash
130
+ pip install -U sentence-transformers
131
+ ```
132
+
133
+ Then you can load this model and run inference.
134
+ ```python
135
+ from sentence_transformers import SentenceTransformer
136
+
137
+ # Download from the 🤗 Hub
138
+ model = SentenceTransformer("sentence_transformers_model_id")
139
+ # Run inference
140
+ sentences = [
141
+ 'Apa saja panduan umum pelaksanaan Program Magang di ITS?',
142
+ 'Mahasiswa dalam melaksanakan magang harus memenuhi ketentuan berikut: 1. Pelaksanaan Program Magang memiliki durasi minimal 1 bulan dan maksimal 6 bulan. 2. Selama Program Magang berlangsung, mahasiswa tidak harus mengajukan cuti. 3. Mahasiswa secara penuh waktu bekerja di lapangan sesuai kesepakatan. 4. Mahasiswa bisa mendapatkan izin untuk kegiatan akademik tertentu dengan kesepakatan pihak Mitra Magang. 5. Mahasiswa dapat mengajukan konversi mata kuliah dengan CPMK yang selaras. 6. Diperlukan pembimbing internal dari Dosen Departemen dan pembimbing lapangan dari Mitra Magang. 7. Sebelum Program Magang, wajib menandatangani perjanjian kerjasama dan nota kesepahaman.',
143
+ '1. Mahasiswa mengurus surat rekomendasi departemen untuk pengajuan magang ke Mitra\n2. Mahasiswa mengajukan permohonan magang ke Mitra\n3. Mitra melakukan seleksi magang\n4. Mahasiswa menerima hasil seleksi magang dari Mitra\n5. Apabila tidak diterima, maka mahasiswa harus mengulang sejak langkah awal\n6. Apabila diterima, Mahasiswa melaporkan ke Departemen\n7. Mahasiswa/Departemen melakukan koordinasi dengan PK2 untuk pengurusan PKS dengan menyertakan Proposal Magang, Surat Rekomendasi Departemen dan Surat Penerimaan Magang\n8. Proses pengurusan PKS (Dapat dilakukan bersamaan dengan pelaksanaan Magang)',
144
+ ]
145
+ embeddings = model.encode(sentences)
146
+ print(embeddings.shape)
147
+ # [3, 384]
148
+
149
+ # Get the similarity scores for the embeddings
150
+ similarities = model.similarity(embeddings, embeddings)
151
+ print(similarities.shape)
152
+ # [3, 3]
153
+ ```
154
+
155
+ <!--
156
+ ### Direct Usage (Transformers)
157
+
158
+ <details><summary>Click to see the direct usage in Transformers</summary>
159
+
160
+ </details>
161
+ -->
162
+
163
+ <!--
164
+ ### Downstream Usage (Sentence Transformers)
165
+
166
+ You can finetune this model on your own dataset.
167
+
168
+ <details><summary>Click to expand</summary>
169
+
170
+ </details>
171
+ -->
172
+
173
+ <!--
174
+ ### Out-of-Scope Use
175
+
176
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
177
+ -->
178
+
179
+ <!--
180
+ ## Bias, Risks and Limitations
181
+
182
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
183
+ -->
184
+
185
+ <!--
186
+ ### Recommendations
187
+
188
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
189
+ -->
190
+
191
+ ## Training Details
192
+
193
+ ### Training Dataset
194
+
195
+ #### Unnamed Dataset
196
+
197
+
198
+ * Size: 704 training samples
199
+ * Columns: <code>sentence_0</code> and <code>sentence_1</code>
200
+ * Approximate statistics based on the first 1000 samples:
201
+ | | sentence_0 | sentence_1 |
202
+ |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
203
+ | type | string | string |
204
+ | details | <ul><li>min: 6 tokens</li><li>mean: 16.22 tokens</li><li>max: 42 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 57.55 tokens</li><li>max: 128 tokens</li></ul> |
205
+ * Samples:
206
+ | sentence_0 | sentence_1 |
207
+ |:-----------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
208
+ | <code>Bagaimana cara menulis dokumen pemerintah dalam daftar pustaka?</code> | <code>Nama pengarang dan tahun publikasi sama dengan penulisan artikel dari jurnal. Judul dokumen dicetak miring/italic. Volume atau nomor (jika ada). Nama penerbit. Kota tempat diterbitkan.</code> |
209
+ | <code>Apa tugas dosen wali dalam pelaksanaan MBKM?</code> | <code>Dosen wali ditugaskan oleh Prodi untuk membuat perencanaan bersama dengan mahasiswa yang akan melaksanakan kegiatan MBKM, melakukan evaluasi terhadap kesesuaian bentuk dan lama pelaksanaan MBKM, serta melakukan penilaian atas rencana, pelaksanaan, dan evaluasi MBKM.</code> |
210
+ | <code>Apa yang dimaksud dengan 'Hak Belajar Tiga Semester di Luar Program Studi'?</code> | <code>Hak Belajar Tiga Semester di Luar Program Studi adalah kebijakan yang memberikan mahasiswa kesempatan untuk satu semester (setara dengan 20 SKS) menempuh pembelajaran di luar program studi pada perguruan tinggi yang sama, dan paling lama dua semester (setara dengan 40 SKS) di program studi yang sama atau berbeda di perguruan tinggi yang berbeda atau di luar perguruan tinggi.</code> |
211
+ * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
212
+ ```json
213
+ {
214
+ "scale": 20.0,
215
+ "similarity_fct": "cos_sim"
216
+ }
217
+ ```
218
+
219
+ ### Training Hyperparameters
220
+ #### Non-Default Hyperparameters
221
+
222
+ - `per_device_train_batch_size`: 16
223
+ - `per_device_eval_batch_size`: 16
224
+ - `num_train_epochs`: 5
225
+ - `multi_dataset_batch_sampler`: round_robin
226
+
227
+ #### All Hyperparameters
228
+ <details><summary>Click to expand</summary>
229
+
230
+ - `overwrite_output_dir`: False
231
+ - `do_predict`: False
232
+ - `prediction_loss_only`: True
233
+ - `per_device_train_batch_size`: 16
234
+ - `per_device_eval_batch_size`: 16
235
+ - `per_gpu_train_batch_size`: None
236
+ - `per_gpu_eval_batch_size`: None
237
+ - `gradient_accumulation_steps`: 1
238
+ - `eval_accumulation_steps`: None
239
+ - `learning_rate`: 5e-05
240
+ - `weight_decay`: 0.0
241
+ - `adam_beta1`: 0.9
242
+ - `adam_beta2`: 0.999
243
+ - `adam_epsilon`: 1e-08
244
+ - `max_grad_norm`: 1
245
+ - `num_train_epochs`: 5
246
+ - `max_steps`: -1
247
+ - `lr_scheduler_type`: linear
248
+ - `lr_scheduler_kwargs`: {}
249
+ - `warmup_ratio`: 0.0
250
+ - `warmup_steps`: 0
251
+ - `log_level`: passive
252
+ - `log_level_replica`: warning
253
+ - `log_on_each_node`: True
254
+ - `logging_nan_inf_filter`: True
255
+ - `save_safetensors`: True
256
+ - `save_on_each_node`: False
257
+ - `save_only_model`: False
258
+ - `no_cuda`: False
259
+ - `use_cpu`: False
260
+ - `use_mps_device`: False
261
+ - `seed`: 42
262
+ - `data_seed`: None
263
+ - `jit_mode_eval`: False
264
+ - `use_ipex`: False
265
+ - `bf16`: False
266
+ - `fp16`: False
267
+ - `fp16_opt_level`: O1
268
+ - `half_precision_backend`: auto
269
+ - `bf16_full_eval`: False
270
+ - `fp16_full_eval`: False
271
+ - `tf32`: None
272
+ - `local_rank`: 0
273
+ - `ddp_backend`: None
274
+ - `tpu_num_cores`: None
275
+ - `tpu_metrics_debug`: False
276
+ - `debug`: []
277
+ - `dataloader_drop_last`: False
278
+ - `dataloader_num_workers`: 0
279
+ - `dataloader_prefetch_factor`: None
280
+ - `past_index`: -1
281
+ - `disable_tqdm`: False
282
+ - `remove_unused_columns`: True
283
+ - `label_names`: None
284
+ - `load_best_model_at_end`: False
285
+ - `ignore_data_skip`: False
286
+ - `fsdp`: []
287
+ - `fsdp_min_num_params`: 0
288
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
289
+ - `fsdp_transformer_layer_cls_to_wrap`: None
290
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}
291
+ - `deepspeed`: None
292
+ - `label_smoothing_factor`: 0.0
293
+ - `optim`: adamw_torch
294
+ - `optim_args`: None
295
+ - `adafactor`: False
296
+ - `group_by_length`: False
297
+ - `length_column_name`: length
298
+ - `ddp_find_unused_parameters`: None
299
+ - `ddp_bucket_cap_mb`: None
300
+ - `ddp_broadcast_buffers`: False
301
+ - `dataloader_pin_memory`: True
302
+ - `dataloader_persistent_workers`: False
303
+ - `skip_memory_metrics`: True
304
+ - `use_legacy_prediction_loop`: False
305
+ - `push_to_hub`: False
306
+ - `resume_from_checkpoint`: None
307
+ - `hub_model_id`: None
308
+ - `hub_strategy`: every_save
309
+ - `hub_private_repo`: False
310
+ - `hub_always_push`: False
311
+ - `gradient_checkpointing`: False
312
+ - `gradient_checkpointing_kwargs`: None
313
+ - `include_inputs_for_metrics`: False
314
+ - `fp16_backend`: auto
315
+ - `push_to_hub_model_id`: None
316
+ - `push_to_hub_organization`: None
317
+ - `mp_parameters`:
318
+ - `auto_find_batch_size`: False
319
+ - `full_determinism`: False
320
+ - `torchdynamo`: None
321
+ - `ray_scope`: last
322
+ - `ddp_timeout`: 1800
323
+ - `torch_compile`: False
324
+ - `torch_compile_backend`: None
325
+ - `torch_compile_mode`: None
326
+ - `dispatch_batches`: None
327
+ - `split_batches`: None
328
+ - `include_tokens_per_second`: False
329
+ - `include_num_input_tokens_seen`: False
330
+ - `neftune_noise_alpha`: None
331
+ - `optim_target_modules`: None
332
+ - `batch_sampler`: batch_sampler
333
+ - `multi_dataset_batch_sampler`: round_robin
334
+
335
+ </details>
336
+
337
+ ### Framework Versions
338
+ - Python: 3.8.10
339
+ - Sentence Transformers: 3.0.1
340
+ - Transformers: 4.39.2
341
+ - PyTorch: 2.3.1+cu121
342
+ - Accelerate: 0.29.3
343
+ - Datasets: 2.19.2
344
+ - Tokenizers: 0.15.2
345
+
346
+ ## Citation
347
+
348
+ ### BibTeX
349
+
350
+ #### Sentence Transformers
351
+ ```bibtex
352
+ @inproceedings{reimers-2019-sentence-bert,
353
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
354
+ author = "Reimers, Nils and Gurevych, Iryna",
355
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
356
+ month = "11",
357
+ year = "2019",
358
+ publisher = "Association for Computational Linguistics",
359
+ url = "https://arxiv.org/abs/1908.10084",
360
+ }
361
+ ```
362
+
363
+ #### MultipleNegativesRankingLoss
364
+ ```bibtex
365
+ @misc{henderson2017efficient,
366
+ title={Efficient Natural Language Response Suggestion for Smart Reply},
367
+ author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
368
+ year={2017},
369
+ eprint={1705.00652},
370
+ archivePrefix={arXiv},
371
+ primaryClass={cs.CL}
372
+ }
373
+ ```
374
+
375
+ <!--
376
+ ## Glossary
377
+
378
+ *Clearly define terms in order to be accessible across audiences.*
379
+ -->
380
+
381
+ <!--
382
+ ## Model Card Authors
383
+
384
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
385
+ -->
386
+
387
+ <!--
388
+ ## Model Card Contact
389
+
390
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
391
+ -->
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2",
3
+ "architectures": [
4
+ "BertModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 384,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 1536,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.39.2",
23
+ "type_vocab_size": 2,
24
+ "use_cache": true,
25
+ "vocab_size": 250037
26
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "3.0.1",
4
+ "transformers": "4.39.2",
5
+ "pytorch": "2.3.1+cu121"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null,
9
+ "similarity_fn_name": null
10
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d322a03f2a6639e63479205d02ed64fafcf97d072a2b490a3f11ca4fee946bcc
3
+ size 470637416
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 128,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "<unk>",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa685fc160bbdbab64058d4fc91b60e62d207e8dc60b9af5c002c5ab946ded00
3
+ size 17083009
tokenizer_config.json ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "250001": {
36
+ "content": "<mask>",
37
+ "lstrip": true,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "<s>",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "<s>",
47
+ "do_lower_case": true,
48
+ "eos_token": "</s>",
49
+ "mask_token": "<mask>",
50
+ "max_length": 128,
51
+ "model_max_length": 128,
52
+ "pad_to_multiple_of": null,
53
+ "pad_token": "<pad>",
54
+ "pad_token_type_id": 0,
55
+ "padding_side": "right",
56
+ "sep_token": "</s>",
57
+ "stride": 0,
58
+ "strip_accents": null,
59
+ "tokenize_chinese_chars": true,
60
+ "tokenizer_class": "BertTokenizer",
61
+ "truncation_side": "right",
62
+ "truncation_strategy": "longest_first",
63
+ "unk_token": "<unk>"
64
+ }
unigram.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da145b5e7700ae40f16691ec32a0b1fdc1ee3298db22a31ea55f57a966c4a65d
3
+ size 14763260