Edit model card

Model Card for Model ID

This is a pre-trained language translation model that aims to create a translation system for English and Swahili lanuages. It is a fine-tuned version of Helsinki-NLP/opus-mt-en-swc on an unknown dataset.

Model Details

  • Transformer architecture used
  • Trained on a 210000 corpus pairs
  • Pre-trained Helsinki-NLP/opus-mt-en-swc
  • 2 models to enforce biderectional translation

Model Description

  • Developed by: Peter Rogendo, Frederick Kioko
  • Model type: Transformer
  • Language(s) (NLP): Transformer, Pandas, Numpy
  • License: Distributed under the MIT License
  • Finetuned from model [Helsinki-NLP/opus-mt-en-swc]: [This pre-trained model was re-trained on a swahili-english sentence pairs that were collected across Kenya. Swahili is the national language and is among the top three of the most spoken language in Africa. The sentences that were used to train this model were 210000 in total.]

Model Sources [optional]

Uses

This translation model is intended to be used in many cases, from language translators, screen assistants, to even in official cases such as translating legal documents.

Direct Use

Use a pipeline as a high-level helper

    from transformers import pipeline
    
    pipe = pipeline("text2text-generation", model="Rogendo/sw-en")

Load model directly

    from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
    
    tokenizer = AutoTokenizer.from_pretrained("Rogendo/sw-en")
    model = AutoModelForSeq2SeqLM.from_pretrained("Rogendo/sw-en")

Downstream Use [optional]

[More Information Needed]

Out-of-Scope Use

[More Information Needed]

Bias, Risks, and Limitations

[More Information Needed]

Recommendations

Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.

How to Get Started with the Model

Use a pipeline as a high-level helper

    from transformers import pipeline
    
    pipe = pipeline("text2text-generation", model="Rogendo/sw-en")

Load model directly

    from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
    
    tokenizer = AutoTokenizer.from_pretrained("Rogendo/sw-en")
    model = AutoModelForSeq2SeqLM.from_pretrained("Rogendo/sw-en")

Training Details

Training Data

curl -X GET
"https://datasets-server.huggingface.co/rows?dataset=Rogendo%2FEnglish-Swahili-Sentence-Pairs&config=default&split=train&offset=0&length=100"

View More https://huggingface.co/datasets/Rogendo/English-Swahili-Sentence-Pairs

Training Procedure

Preprocessing [optional]

[More Information Needed]

Training Hyperparameters

  • Training regime: [More Information Needed]

Speeds, Sizes, Times [optional]

[More Information Needed]

Evaluation

Testing Data, Factors & Metrics

Testing Data

[More Information Needed]

Factors

[More Information Needed]

Metrics

[More Information Needed]

Results

[More Information Needed]

Summary

Model Examination [optional]

[More Information Needed]

Environmental Impact

Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).

  • Hardware Type: [More Information Needed]
  • Hours used: [More Information Needed]
  • Cloud Provider: [More Information Needed]
  • Compute Region: [More Information Needed]
  • Carbon Emitted: [More Information Needed]

Technical Specifications [optional]

Model Architecture and Objective

[More Information Needed]

Compute Infrastructure

[More Information Needed]

Hardware

[More Information Needed]

Software

[More Information Needed]

Citation [optional]

BibTeX:

[More Information Needed]

APA:

[More Information Needed]

Glossary [optional]

Model Card Authors [optional]

Peter Rogendo

Model Card Contact

progendo@kabarak.ac.ke

Downloads last month
24
Safetensors
Model size
74.4M params
Tensor type
F32
ยท

Space using Rogendo/sw-en 1