Edit model card

SwahiliGPT

This is a Mistral model that has been fine-tuned on the Wikipedia Swahili dataset.

Usage

# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("mwitiderrick/SwahiliGPT_v0.1")
model = AutoModelForCausalLM.from_pretrained("mwitiderrick/SwahiliGPT_v0.1", device_map="auto")
inputs = tokenizer("Hapo zamani za kale", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=200,  do_sample=True, repetition_penalty=1.1)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))


"""
Hapo zamani za kale katika historia ya jamii, ambavyo sehemu moja hutazama historia ile inayopendekezwa au inayojulikana, na sehemu nyingine inafanya history  ambalai hainajulikana. 
Utaifishaji unaleta utata kwanza mambo ya karne zilizoandamana, na seconda matokeo yanatokana na vipitio vya maisha muhimu ambavyo haivyo vitakuva mahitaji katika jamii hiyo (hunajua wakiweka mitindo katakatani). Ni kinyume kingine kwamba kuna sifa ambayo umechukizwa vitu hivi vilitengenezwa zaidi.

Katika Afrika Magharibi, historia huitwa ngan
"""
Downloads last month
10
Safetensors
Model size
7.24B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for mwitiderrick/SwahiliGPT_v0.1

Finetuned
(680)
this model

Dataset used to train mwitiderrick/SwahiliGPT_v0.1