Text Generation
Transformers
PyTorch
Igbo
Yoruba
English
llama
conversational
Inference Endpoints
text-generation-inference
Edit model card

Model Name: Multilingual LLM for English, Igbo, and Yoruba Translation

This model specializes in understanding and generating text across English, Igbo, and Yoruba, trained on curated datasets including "tiny_stories" and "dolly_hhrlhf." Utilizing advanced techniques to ensure efficiency and accuracy, the model supports a wide range of applications from translation to content creation.

Key Features:

  • Trilingual Support: Seamlessly processes and generates content in English, Igbo, and Yoruba, promoting linguistic diversity and accessibility.
  • Custom Training Approach: Employs a tailored training setup, leveraging specific prompts and responses to enhance model performance on relevant tasks.
  • High Sequence Capability: Handles extensive text inputs (up to 3453 tokens), making it suitable for detailed narratives and complex translation tasks.
  • Efficiency Optimizations: Incorporates strategies such as dataset packing, reducing computational demands while maintaining high-quality output.

Applications:

  • Translation Services: Offers precise, context-aware translations, bridging communication gaps between English, Igbo, and Yoruba speakers.
  • Content Generation: Generates culturally and linguistically nuanced content, catering to a diverse audience.
  • Educational Tools: Assists in language learning and preservation, providing resources in underrepresented languages.

Future Directions:

  • Further refinement with diverse text sources to enhance understanding and generation capabilities.
  • Expansion to additional languages, supporting broader multilingual communication.

This model represents a step towards more inclusive language technologies, recognizing the importance of language diversity in global communication.

Training Loss:

  • Below is the training loss.
Step Training Loss
200 1.490900
400 1.375600
600 1.304100
800 1.198700
1000 1.228200
1125 1.226600

About the Creators

Christopher Ibe and Okezie Okoye continue to lead Hypa AI towards new frontiers in AI translation. Their dedication to leveraging advanced AI for genuine understanding and connection across language barriers is what sets Hypa AI apart in the field of artificial intelligence.

Hypa AI remains steadfast in its mission to pioneer intelligent solutions that are not just technologically advanced but are also culturally aware, ensuring that the future of AI is as diverse and inclusive as the world it serves.

AfroVoices, a subsidiary of Hypa AI, is dedicated to amplifying African voices, languages, and cultures in the intelligence age. Focused on bridging the digital representation gap, AfroVoices curates datasets and resources for African languages, promoting inclusivity and cultural appreciation in AI technologies. Their mission goes beyond technological innovation, aiming to celebrate the richness of African linguistic diversity on a global stage.


Model Card for Model ID

Model Details

Model Description

This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.

  • Developed by: [More Information Needed]
  • Funded by [optional]: [More Information Needed]
  • Shared by [optional]: [More Information Needed]
  • Model type: [More Information Needed]
  • Language(s) (NLP): [More Information Needed]
  • License: [More Information Needed]
  • Finetuned from model [optional]: [More Information Needed]

Model Sources [optional]

  • Repository: [More Information Needed]
  • Paper [optional]: [More Information Needed]
  • Demo [optional]: [More Information Needed]

Uses

Direct Use

[More Information Needed]

Downstream Use [optional]

[More Information Needed]

Out-of-Scope Use

[More Information Needed]

Bias, Risks, and Limitations

[More Information Needed]

Recommendations

Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.

How to Get Started with the Model

Use the code below to get started with the model.

[More Information Needed]

Training Details

Training Data

[More Information Needed]

Training Procedure

Preprocessing [optional]

[More Information Needed]

Training Hyperparameters

  • Training regime: [More Information Needed]

Speeds, Sizes, Times [optional]

[More Information Needed]

Evaluation

Testing Data, Factors & Metrics

Testing Data

[More Information Needed]

Factors

[More Information Needed]

Metrics

[More Information Needed]

Results

[More Information Needed]

Summary

Model Examination [optional]

[More Information Needed]

Environmental Impact

Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).

  • Hardware Type: [More Information Needed]
  • Hours used: [More Information Needed]
  • Cloud Provider: [More Information Needed]
  • Compute Region: [More Information Needed]
  • Carbon Emitted: [More Information Needed]

Technical Specifications [optional]

Model Architecture and Objective

[More Information Needed]

Compute Infrastructure

[More Information Needed]

Hardware

[More Information Needed]

Software

[More Information Needed]

Citation [optional]

BibTeX:

[More Information Needed]

APA:

[More Information Needed]

Glossary [optional]

[More Information Needed]

More Information [optional]

[More Information Needed]

Model Card Authors [optional]

[More Information Needed]

Model Card Contact

[More Information Needed]

Downloads last month
39

Datasets used to train ccibeekeoc42/Llama3-8b-chat-SFT-2024-04-20