Edit model card

T5 for belarusian language

model image

This model is based on T5-small with sequence length equal 128 tokens. Model trained from scratch on RTX 3090 24GB.

Supported tasks:

  • translation BE to RU: <extra_id_1>
  • translation BE to EN: <extra_id_2>
  • translation RU to BE: <extra_id_3>
  • translation RU to EN: <extra_id_5>
  • translation EN to BE: <extra_id_6>
  • translation EN to RU: <extra_id_7>

Metrics:

How to Get Started with the Model

Click to expand
from transformers import T5TokenizerFast, T5ForConditionalGeneration

tokenizer = T5TokenizerFast.from_pretrained("WelfCrozzo/T5-L128-belarusian")
model = T5ForConditionalGeneration.from_pretrained("WelfCrozzo/T5-L128-belarusian")

x = tokenizer.encode('<extra_id_1>да зорак праз цяжкасці', return_tensors='pt')

result = model.generate(x, return_dict_in_generate=True, output_scores=True,max_length=128)
print(tokenizer.decode(result["sequences"][0]))

References

Downloads last month
2
Safetensors
Model size
20.7M params
Tensor type
F32
·

Dataset used to train WelfCrozzo/T5-L128-belarusian