This model was trained using HuggingFace's Flax framework and is part of the JAX/Flax Community Week organized by HuggingFace. All training was done on a TPUv3-8 VM sponsored by the Google Cloud team.
from transformers import AutoTokenizer, AutoModelWithLMHead tokenizer = AutoTokenizer.from_pretrained("flax-community/gpt2-swahili") model = AutoModelWithLMHead.from_pretrained("flax-community/gpt2-swahili") print(round((model.num_parameters())/(1000*1000)),"Million Parameters") 124 Million Parameters
This model was trained on Swahili Safi
For more details and Demo please check HF Swahili Space
- Downloads last month