SentenceTransformer Fine-Tuned for Amharic Retrieval

This model is a sentence-transformers model finetuned on Amharic QA triplets. It maps sentences and paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.

Model Details

  • Model Type: Sentence Transformer
  • Base Model: sentence-transformers/paraphrase-xlm-r-multilingual-v1
  • Training Task: Triplet loss with Matryoshka loss
  • Language: Amharic
  • Maximum Sequence Length: 128 tokens
  • Output Dimensionality: 768 dimensions
  • Similarity Function: Cosine Similarity

Training Overview

  • Training Data: Custom Amharic QA triplets (with positive and negative examples)
  • Training Strategy:
    The model was finetuned using a combination of triplet loss and a Matryoshka loss, with evaluation performed using a TripletEvaluator.
  • Hyperparameters:
    • Epochs: 3
    • Batch Size: 16
    • Learning Rate: 1e-6
    • Warmup Ratio: 0.08
    • Weight Decay: 0.05

Evaluation

The model was evaluated on a held-out test set using cosine similarity as the metric:

Metric Value
Cosine Accuracy 0.875

Usage

To use the model in your own project:

  1. Install Sentence Transformers:

    pip install -U sentence-transformers
    
  2. Load the Model:

    from sentence_transformers import SentenceTransformer
    
    model = SentenceTransformer("abdulmunimjemal/xlm-r-retrieval-am-v5")
    sentences = [
        "αˆ°αˆ›α‹­ αˆαŠ• αŠ α‹­αŠα‰΅ α‰€αˆˆαˆ αŠα‹?",
        "αˆ°αˆ›α‹­ αˆ°αˆ›α‹«α‹Š α‰€αˆˆαˆ αŠ αˆˆα‹α’" ,
        "αŠ₯αŠ” ምሳ αŠ₯αŠ•αŒ€αˆ« α‰ αˆ‹αˆα’" ,
        "α‰£αˆ•αˆ­ αˆαŠ• αŠ α‹­αŠα‰΅ α‰€αˆˆαˆ αŠα‹?",
        "αŠ α‹¨αˆ­ α‰ αˆα‹΅αˆ­ α‹™αˆͺα‹« α‹«αˆˆ αŠα‹α’"
    ]
    embeddings = model.encode(sentences)
    print(embeddings.shape)  # Expected output: (5, 768)
    
  3. Compute Similarity:

    from sklearn.metrics.pairwise import cosine_similarity
    similarities = cosine_similarity(embeddings, embeddings)
    print(similarities.shape)  # Expected output: (5, 5)
    

Model Architecture

Below is an outline of the model architecture:

SentenceTransformer(
  (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: XLMRobertaModel 
  (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_mean_tokens': True, ...})
)

Training Environment

  • Python: 3.11.11
  • Sentence Transformers: 3.3.1
  • Transformers: 4.47.1
  • PyTorch: 2.5.1+cu124
  • Accelerate: 1.2.1
  • Datasets: 3.2.0
  • Tokenizers: 0.21.0

Citation

If you use this model in your research, please cite it appropriately.

@misc{your_model,
  title = {SentenceTransformer Fine-Tuned for Amharic Retrieval},
  author = {Abdulmunim J. Jemal},
  year = {2025},
  howpublished = {Hugging Face Model Hub, \url{https://huggingface.co/abdulmunimjemal/xlm-r-retrieval-am-v1}}
}
Downloads last month
32
Safetensors
Model size
278M params
Tensor type
F32
Β·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Evaluation results