gpt2-tigrinya-small / README.md
luel's picture
Update README.md
895ff25 verified
|
raw
history blame
851 Bytes
metadata
language: ti
tags:
  - tigrinya
  - gpt2
  - text-generation
library_name: transformers
pipeline_tag: text-generation

GPT-2 for Tigrinya Language

This repository contains a GPT-2 model trained from scratch on Tigrinya text data. The model was trained using the Hugging Face Transformers library.

Model Details

  • Model Type: GPT-2
  • Language: Tigrinya
  • Vocabulary Size: 16000
  • Maximum Length: 128

Training Details

  • Number of Epochs: 12
  • Batch Size: 1 (with gradient accumulation steps of 4)
  • Learning Rate: 5e-4

Dataset Statistics

  • Total number of words: 16061839
  • Total number of unique words: 458901

Usage

from transformers import pipeline

# Load the model
generator = pipeline('text-generation', model='luel/gpt2-tigrinya-small')

# Generate text
text = generator("ትግራይ", max_length=60)
print(text)