arabic-nlp
Collection
Arabic focused finetuned models • 3 items • Updated
How to use RamzyBakir/arabic-gte-multilingual-embed-medium with sentence-transformers:
from sentence_transformers import SentenceTransformer
model = SentenceTransformer("RamzyBakir/arabic-gte-multilingual-embed-medium", trust_remote_code=True)
sentences = [
"So our minds are filling in those other 275 words.",
"لذا فإن عقولنا تكون مشغولة وممتلئة ب 275 كلمة كتلك.",
"هل يمكنك الحمل قبل يوم واحد من الإباضة؟",
"وهذه المقاطع تكوّن كلمات ."
]
embeddings = model.encode(sentences)
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
# [4, 4]This is a sentence-transformers model finetuned from Alibaba-NLP/gte-multilingual-base on the silma-arabic-triplets-dataset-v1.0 dataset. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for retrieval.
SentenceTransformer(
(0): Transformer({'transformer_task': 'feature-extraction', 'modality_config': {'text': {'method': 'forward', 'method_output_name': 'last_hidden_state'}}, 'module_output_name': 'token_embeddings', 'architecture': 'NewModel'})
(1): Pooling({'embedding_dimension': 768, 'pooling_mode': 'cls', 'include_prompt': True})
(2): Normalize({})
)
First install the Sentence Transformers library:
pip install -U sentence-transformers
Then you can load this model and run inference.
from sentence_transformers import SentenceTransformer
# Download from the 🤗 Hub
model = SentenceTransformer("RamzyBakir/arabic-gte-multilingual-embed-medium")
# Run inference
sentences = [
'هم نوع من مثل القطط.',
'إنهم يشبهون القطط',
'إنسان مع قطة',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 768]
# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities)
# tensor([[1.0000, 0.8787, 0.6596],
# [0.8787, 1.0000, 0.7666],
# [0.6596, 0.7666, 1.0000]])
arabic-triplet-fast and arabic-triplet-eval-fullTripletEvaluator| Metric | arabic-triplet-fast | arabic-triplet-eval-full |
|---|---|---|
| cosine_accuracy | 0.9804 | 0.9824 |
anchor, positive, and negative| anchor | positive | negative | |
|---|---|---|---|
| type | string | string | string |
| details |
|
|
|
| anchor | positive | negative |
|---|---|---|
They think the web is a force for good, and most don’t want governments to regulate it. |
وهم يعتقدون أن شبكة الويب هي قوة تستخدم للخير، ولا يرغب أغلبهم في أن تقوم الحكومات بتنظيم تلك الشبكة. |
أصبحت الحكومة أكثر حساسية لتأثير الإنترنت على السياسة الداخلية وسنت قوانين تزيد من سلطتها لتنظيم هذا القطاع. |
طفل صغير يرتدي نظارات زرقاء يجلس على طوف في بركة سباحة. |
طفل صغير يجلس على طوف في حمام سباحة ويرتدي نظارات زرقاء. |
امرأة ترقص بعنف على خشبة المسرح |
امرأة وطفل يسيرون على الرصيف المغطى بالأوراق متجهين نحو شخصين يركبان خيول. |
سيدة وطفل يسيرون على الرصيف باتجاه الخيول |
امرأة وطفلان يسيران في حديقة. |
CachedMultipleNegativesRankingLoss with these parameters:{
"scale": 50.0,
"similarity_fct": "cos_sim",
"mini_batch_size": 64,
"gather_across_devices": false,
"directions": [
"query_to_doc"
],
"partition_mode": "joint",
"hardness_mode": null,
"hardness_strength": 0.0
}
anchor, positive, and negative| anchor | positive | negative | |
|---|---|---|---|
| type | string | string | string |
| details |
|
|
|
| anchor | positive | negative |
|---|---|---|
It is typically called the Constitution of the Fifth Republic, and replaced that of the Fourth Republic dating from 1946. |
ويسمى عادة دستور الجمهورية الخامسة، وحل محل دستور الجمهورية الرابع الذي يعود تاريخه إلى عام 1946. |
في نوفمبر 1946، اعتمدت الجمعية الوطنية أول دستور للجمهورية. |
ما هو أفضل يوم في حياتك؟ |
أي يوم في حياتك اعتبرته أفضل يوم في حياتك؟ |
ما هي وصفات الملفات؟ |
She married French researcher Jean Ghata. |
تزوجت من الباحث الفرنسي جان غاتا. |
ثم انتقلوا إلى فرنسا حيث تزوجوا. |
CachedMultipleNegativesRankingLoss with these parameters:{
"scale": 50.0,
"similarity_fct": "cos_sim",
"mini_batch_size": 64,
"gather_across_devices": false,
"directions": [
"query_to_doc"
],
"partition_mode": "joint",
"hardness_mode": null,
"hardness_strength": 0.0
}
per_device_train_batch_size: 128per_device_eval_batch_size: 256learning_rate: 2e-05max_steps: 15842lr_scheduler_type: cosine_with_min_lrlr_scheduler_kwargs: {'min_lr': 1e-06}warmup_ratio: 0.05bf16: Truedataloader_num_workers: 4remove_unused_columns: Falseload_best_model_at_end: Truebatch_sampler: no_duplicatesoverwrite_output_dir: Falsedo_predict: Falseprediction_loss_only: Trueper_device_train_batch_size: 128per_device_eval_batch_size: 256per_gpu_train_batch_size: Noneper_gpu_eval_batch_size: Nonegradient_accumulation_steps: 1eval_accumulation_steps: Nonetorch_empty_cache_steps: Nonelearning_rate: 2e-05weight_decay: 0.0adam_beta1: 0.9adam_beta2: 0.999adam_epsilon: 1e-08max_grad_norm: 1.0num_train_epochs: 3.0max_steps: 15842lr_scheduler_type: cosine_with_min_lrlr_scheduler_kwargs: {'min_lr': 1e-06}warmup_ratio: 0.05warmup_steps: 0log_level: passivelog_level_replica: warninglog_on_each_node: Truelogging_nan_inf_filter: Truesave_safetensors: Truesave_on_each_node: Falsesave_only_model: Falserestore_callback_states_from_checkpoint: Falseno_cuda: Falseuse_cpu: Falseuse_mps_device: Falseseed: 42data_seed: Nonejit_mode_eval: Falseuse_ipex: Falsebf16: Truefp16: Falsefp16_opt_level: O1half_precision_backend: autobf16_full_eval: Falsefp16_full_eval: Falsetf32: Nonelocal_rank: 0ddp_backend: Nonetpu_num_cores: Nonetpu_metrics_debug: Falsedebug: []dataloader_drop_last: Falsedataloader_num_workers: 4dataloader_prefetch_factor: Nonepast_index: -1disable_tqdm: Falseremove_unused_columns: Falselabel_names: Noneload_best_model_at_end: Trueignore_data_skip: Falsefsdp: []fsdp_min_num_params: 0fsdp_config: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}fsdp_transformer_layer_cls_to_wrap: Noneaccelerator_config: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}parallelism_config: Nonedeepspeed: Nonelabel_smoothing_factor: 0.0optim: adamw_torch_fusedoptim_args: Noneadafactor: Falsegroup_by_length: Falselength_column_name: lengthddp_find_unused_parameters: Noneddp_bucket_cap_mb: Noneddp_broadcast_buffers: Falsedataloader_pin_memory: Truedataloader_persistent_workers: Falseskip_memory_metrics: Trueuse_legacy_prediction_loop: Falsepush_to_hub: Falseresume_from_checkpoint: Nonehub_model_id: Nonehub_strategy: every_savehub_private_repo: Nonehub_always_push: Falsehub_revision: Nonegradient_checkpointing: Falsegradient_checkpointing_kwargs: Noneinclude_inputs_for_metrics: Falseinclude_for_metrics: []eval_do_concat_batches: Truefp16_backend: autopush_to_hub_model_id: Nonepush_to_hub_organization: Nonemp_parameters: auto_find_batch_size: Falsefull_determinism: Falsetorchdynamo: Noneray_scope: lastddp_timeout: 1800torch_compile: Falsetorch_compile_backend: Nonetorch_compile_mode: Noneinclude_tokens_per_second: Falseinclude_num_input_tokens_seen: Falseneftune_noise_alpha: Noneoptim_target_modules: Nonebatch_eval_metrics: Falseeval_on_start: Falseuse_liger_kernel: Falseliger_kernel_config: Noneeval_use_gather_object: Falseaverage_tokens_across_devices: Falseprompts: Nonebatch_sampler: no_duplicatesmulti_dataset_batch_sampler: proportionalrouter_mapping: {}learning_rate_mapping: {}| Epoch | Step | Training Loss | Validation Loss | arabic-triplet-fast_cosine_accuracy | arabic-triplet-eval-full_cosine_accuracy |
|---|---|---|---|---|---|
| 0.0001 | 1 | 0.5997 | - | - | - |
| 0.0063 | 100 | 0.4092 | - | - | - |
| 0.0126 | 200 | 0.3724 | - | - | - |
| 0.0189 | 300 | 0.3321 | - | - | - |
| 0.0252 | 400 | 0.3078 | - | - | - |
| 0.0316 | 500 | 0.2871 | - | - | - |
| 0.0379 | 600 | 0.2782 | - | - | - |
| 0.0442 | 700 | 0.2669 | - | - | - |
| 0.0505 | 800 | 0.2679 | - | - | - |
| 0.0568 | 900 | 0.2628 | - | - | - |
| 0.0631 | 1000 | 0.252 | - | - | - |
| 0.0694 | 1100 | 0.2424 | - | - | - |
| 0.0757 | 1200 | 0.2474 | - | - | - |
| 0.0821 | 1300 | 0.2321 | - | - | - |
| 0.0884 | 1400 | 0.2484 | - | - | - |
| 0.0947 | 1500 | 0.2309 | - | - | - |
| 0.1000 | 1584 | - | 0.3115 | 0.9704 | - |
| 0.1010 | 1600 | 0.2291 | - | - | - |
| 0.1073 | 1700 | 0.2242 | - | - | - |
| 0.1136 | 1800 | 0.2142 | - | - | - |
| 0.1199 | 1900 | 0.209 | - | - | - |
| 0.1262 | 2000 | 0.2259 | - | - | - |
| 0.1326 | 2100 | 0.2169 | - | - | - |
| 0.1389 | 2200 | 0.2076 | - | - | - |
| 0.1452 | 2300 | 0.2129 | - | - | - |
| 0.1515 | 2400 | 0.2041 | - | - | - |
| 0.1578 | 2500 | 0.204 | - | - | - |
| 0.1641 | 2600 | 0.2173 | - | - | - |
| 0.1704 | 2700 | 0.2086 | - | - | - |
| 0.1767 | 2800 | 0.2032 | - | - | - |
| 0.1830 | 2900 | 0.2101 | - | - | - |
| 0.1894 | 3000 | 0.2026 | - | - | - |
| 0.1957 | 3100 | 0.1977 | - | - | - |
| 0.2000 | 3168 | - | 0.2913 | 0.9713 | - |
| 0.2020 | 3200 | 0.1975 | - | - | - |
| 0.2083 | 3300 | 0.2019 | - | - | - |
| 0.2146 | 3400 | 0.1999 | - | - | - |
| 0.2209 | 3500 | 0.191 | - | - | - |
| 0.2272 | 3600 | 0.1996 | - | - | - |
| 0.2335 | 3700 | 0.201 | - | - | - |
| 0.2399 | 3800 | 0.1976 | - | - | - |
| 0.2462 | 3900 | 0.1963 | - | - | - |
| 0.2525 | 4000 | 0.1903 | - | - | - |
| 0.2588 | 4100 | 0.1826 | - | - | - |
| 0.2651 | 4200 | 0.1879 | - | - | - |
| 0.2714 | 4300 | 0.1764 | - | - | - |
| 0.2777 | 4400 | 0.1864 | - | - | - |
| 0.2840 | 4500 | 0.1909 | - | - | - |
| 0.2903 | 4600 | 0.1803 | - | - | - |
| 0.2967 | 4700 | 0.1819 | - | - | - |
| 0.2999 | 4752 | - | 0.2619 | 0.9762 | - |
| 0.3030 | 4800 | 0.187 | - | - | - |
| 0.3093 | 4900 | 0.1904 | - | - | - |
| 0.3156 | 5000 | 0.1899 | - | - | - |
| 0.3219 | 5100 | 0.1764 | - | - | - |
| 0.3282 | 5200 | 0.1828 | - | - | - |
| 0.3345 | 5300 | 0.1725 | - | - | - |
| 0.3408 | 5400 | 0.1674 | - | - | - |
| 0.3472 | 5500 | 0.1757 | - | - | - |
| 0.3535 | 5600 | 0.166 | - | - | - |
| 0.3598 | 5700 | 0.178 | - | - | - |
| 0.3661 | 5800 | 0.1765 | - | - | - |
| 0.3724 | 5900 | 0.1677 | - | - | - |
| 0.3787 | 6000 | 0.1653 | - | - | - |
| 0.3850 | 6100 | 0.176 | - | - | - |
| 0.3913 | 6200 | 0.1533 | - | - | - |
| 0.3977 | 6300 | 0.1622 | - | - | - |
| 0.3999 | 6336 | - | 0.2459 | 0.9771 | - |
| 0.4040 | 6400 | 0.1741 | - | - | - |
| 0.4103 | 6500 | 0.1624 | - | - | - |
| 0.4166 | 6600 | 0.1639 | - | - | - |
| 0.4229 | 6700 | 0.1674 | - | - | - |
| 0.4292 | 6800 | 0.1665 | - | - | - |
| 0.4355 | 6900 | 0.1679 | - | - | - |
| 0.4418 | 7000 | 0.1611 | - | - | - |
| 0.4481 | 7100 | 0.1661 | - | - | - |
| 0.4545 | 7200 | 0.1684 | - | - | - |
| 0.4608 | 7300 | 0.1674 | - | - | - |
| 0.4671 | 7400 | 0.1746 | - | - | - |
| 0.4734 | 7500 | 0.1684 | - | - | - |
| 0.4797 | 7600 | 0.1667 | - | - | - |
| 0.4860 | 7700 | 0.1605 | - | - | - |
| 0.4923 | 7800 | 0.1537 | - | - | - |
| 0.4986 | 7900 | 0.171 | - | - | - |
| 0.4999 | 7920 | - | 0.2387 | 0.9767 | - |
| 0.5050 | 8000 | 0.1587 | - | - | - |
| 0.5113 | 8100 | 0.1623 | - | - | - |
| 0.5176 | 8200 | 0.1704 | - | - | - |
| 0.5239 | 8300 | 0.1575 | - | - | - |
| 0.5302 | 8400 | 0.1671 | - | - | - |
| 0.5365 | 8500 | 0.1608 | - | - | - |
| 0.5428 | 8600 | 0.1537 | - | - | - |
| 0.5491 | 8700 | 0.1568 | - | - | - |
| 0.5555 | 8800 | 0.1582 | - | - | - |
| 0.5618 | 8900 | 0.1598 | - | - | - |
| 0.5681 | 9000 | 0.1613 | - | - | - |
| 0.5744 | 9100 | 0.1628 | - | - | - |
| 0.5807 | 9200 | 0.1507 | - | - | - |
| 0.5870 | 9300 | 0.148 | - | - | - |
| 0.5933 | 9400 | 0.1573 | - | - | - |
| 0.5996 | 9500 | 0.147 | - | - | - |
| 0.5999 | 9504 | - | 0.2270 | 0.9788 | - |
| 0.6059 | 9600 | 0.1502 | - | - | - |
| 0.6123 | 9700 | 0.1445 | - | - | - |
| 0.6186 | 9800 | 0.1534 | - | - | - |
| 0.6249 | 9900 | 0.1544 | - | - | - |
| 0.6312 | 10000 | 0.1509 | - | - | - |
| 0.6375 | 10100 | 0.1599 | - | - | - |
| 0.6438 | 10200 | 0.1579 | - | - | - |
| 0.6501 | 10300 | 0.1525 | - | - | - |
| 0.6564 | 10400 | 0.1371 | - | - | - |
| 0.6628 | 10500 | 0.1456 | - | - | - |
| 0.6691 | 10600 | 0.148 | - | - | - |
| 0.6754 | 10700 | 0.1472 | - | - | - |
| 0.6817 | 10800 | 0.1448 | - | - | - |
| 0.6880 | 10900 | 0.1488 | - | - | - |
| 0.6943 | 11000 | 0.1589 | - | - | - |
| 0.6999 | 11088 | - | 0.2218 | 0.9799 | - |
| 0.7006 | 11100 | 0.1464 | - | - | - |
| 0.7069 | 11200 | 0.1391 | - | - | - |
| 0.7132 | 11300 | 0.1489 | - | - | - |
| 0.7196 | 11400 | 0.1492 | - | - | - |
| 0.7259 | 11500 | 0.1561 | - | - | - |
| 0.7322 | 11600 | 0.1498 | - | - | - |
| 0.7385 | 11700 | 0.1553 | - | - | - |
| 0.7448 | 11800 | 0.1485 | - | - | - |
| 0.7511 | 11900 | 0.1432 | - | - | - |
| 0.7574 | 12000 | 0.1385 | - | - | - |
| 0.7637 | 12100 | 0.1497 | - | - | - |
| 0.7701 | 12200 | 0.145 | - | - | - |
| 0.7764 | 12300 | 0.1354 | - | - | - |
| 0.7827 | 12400 | 0.1345 | - | - | - |
| 0.7890 | 12500 | 0.1472 | - | - | - |
| 0.7953 | 12600 | 0.141 | - | - | - |
| 0.7998 | 12672 | - | 0.2167 | 0.9802 | - |
| 0.8016 | 12700 | 0.1376 | - | - | - |
| 0.8079 | 12800 | 0.1332 | - | - | - |
| 0.8142 | 12900 | 0.1469 | - | - | - |
| 0.8206 | 13000 | 0.142 | - | - | - |
| 0.8269 | 13100 | 0.1391 | - | - | - |
| 0.8332 | 13200 | 0.1512 | - | - | - |
| 0.8395 | 13300 | 0.1467 | - | - | - |
| 0.8458 | 13400 | 0.1485 | - | - | - |
| 0.8521 | 13500 | 0.1485 | - | - | - |
| 0.8584 | 13600 | 0.1412 | - | - | - |
| 0.8647 | 13700 | 0.1482 | - | - | - |
| 0.8710 | 13800 | 0.1532 | - | - | - |
| 0.8774 | 13900 | 0.1402 | - | - | - |
| 0.8837 | 14000 | 0.136 | - | - | - |
| 0.8900 | 14100 | 0.1416 | - | - | - |
| 0.8963 | 14200 | 0.1427 | - | - | - |
| 0.8998 | 14256 | - | 0.2136 | 0.9800 | - |
| 0.9026 | 14300 | 0.1496 | - | - | - |
| 0.9089 | 14400 | 0.1415 | - | - | - |
| 0.9152 | 14500 | 0.1395 | - | - | - |
| 0.9215 | 14600 | 0.1367 | - | - | - |
| 0.9279 | 14700 | 0.1424 | - | - | - |
| 0.9342 | 14800 | 0.1421 | - | - | - |
| 0.9405 | 14900 | 0.1312 | - | - | - |
| 0.9468 | 15000 | 0.1427 | - | - | - |
| 0.9531 | 15100 | 0.1421 | - | - | - |
| 0.9594 | 15200 | 0.1347 | - | - | - |
| 0.9657 | 15300 | 0.141 | - | - | - |
| 0.9720 | 15400 | 0.144 | - | - | - |
| 0.9784 | 15500 | 0.1417 | - | - | - |
| 0.9847 | 15600 | 0.1416 | - | - | - |
| 0.9910 | 15700 | 0.1356 | - | - | - |
| 0.9973 | 15800 | 0.1403 | - | - | - |
| 0.9998 | 15840 | - | 0.2124 | 0.9804 | - |
| -1 | -1 | - | - | - | 0.9824 |
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "https://arxiv.org/abs/1908.10084",
}
@misc{gao2021scaling,
title={Scaling Deep Contrastive Learning Batch Size under Memory Limited Setup},
author={Luyu Gao and Yunyi Zhang and Jiawei Han and Jamie Callan},
year={2021},
eprint={2101.06983},
archivePrefix={arXiv},
primaryClass={cs.LG}
}
Base model
Alibaba-NLP/gte-multilingual-base