GPT-2 for Tigrinya Language

This repository contains a GPT-2 model trained from scratch on Tigrinya text data. The model was trained using the Hugging Face Transformers library.

Model Details

  • Model Type: GPT-2
  • Language: Tigrinya
  • Vocabulary Size: 16000
  • Maximum Length: 128

Training Details

  • Number of Epochs: 12
  • Batch Size: 1 (with gradient accumulation steps of 4)
  • Learning Rate: 5e-4

Dataset Statistics

  • Total number of words: 16061839
  • Total number of unique words: 458901

Usage

from transformers import pipeline

# Load the model
generator = pipeline('text-generation', model='luel/gpt2-tigrinya-small')

# Generate text
text = generator("ትግራይ", max_length=60)
print(text)
Downloads last month
154
Safetensors
Model size
16.8M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Collection including luel/gpt2-tigrinya-small