Text Generation
Transformers
PyTorch
Polish
English
llama
text-generation-inference
Inference Endpoints
Edit model card

image/png

polka-1.1b

polka-1.1b takes the TinyLlama-1.1B model and enhances it by continuing pretraining on an additional 5.7 billion Polish tokens, primarily sourced from the MADLAD-400 dataset. The tokens were sampled in a 10:1 ratio between Polish and English shards using DSIR. Furthermore, Polka extends the TinyLlama tokenizer's vocabulary to 43,882 tokens, improving its efficiency for generating Polish text.

The training took 680 GPU hours on a single 8 x RTX 4090 machine with DeepSpeed ZeRO-2.

Context size: 2,048 tokens.

Notes

This base model was initially developed as the foundation for instruction tuning, which resulted in polka-1.1b-chat. Nonetheless, I'm sharing it with the community because I see potential value in its combination of relatively good performance and an efficient bilingual tokenizer.

The model is capable of producing coherent Polish text, but due to its size, it is likely to suffer from hallucination.

Evaluation

Performed by OPI-PG, the authors of Qra models.

PolEval-2018

ModelPerplexity
English models
meta-llama/Llama-2-7b-hf24.3
meta-llama/Llama-2-13b-hf21.4
mistralai/Mistral-7B-v0.121.4
TinyLlama/TinyLlama-1.1B40.4
Polish models
sdadas/polish-gpt2-small134.4
sdadas/polish-gpt2-medium100.8
sdadas/polish-gpt2-large93.2
sdadas/polish-gpt2-xl94.1
Azurro/APT3-275M-Base129.8
Azurro/APT3-500M-Base153.1
Azurro/APT3-1B-Base106.8
eryk-mazus/polka-1.1b18.1
szymonrucinski/Curie-7B-v113.5
OPI-PG/Qra-1b14.7

Long documents (2024)

Currently, LLMs support contexts of thousands of tokens. Their practical applications usually also involve processing long documents. Therefore, evaluating perplexity on a sentence-based dataset such as PolEval-2018 may not be meaningful. Additionally, the PolEval corpus has been publicly available on the internet for the past few years, which raises the possibility that for some models the training sets have been contaminated by this data. For this reason, we have prepared a new collection consisting of long papers published exclusively in 2024, which will allow us to more reliably test the perplexities of the models on new knowledge that was not available to them at the time of training. The corpus consists of 5,000 documents ranging from several hundred to about 20,000 tokens. Half of the set consists of press texts from Polish news portals from February 2024, the other half are scientific articles published since January 2024. Most of the documents exceed the context size of the evaluated models. To calculate perplexity for these documents, we divided them into chunks of size equal to the model's context length with a stride of 512 tokens, following this example.

ModelContextPerplexity
English models
meta-llama/Llama-2-7b-hf40965.9
meta-llama/Llama-2-13b-hf40965.3
mistralai/Mistral-7B-v0.140964.9
TinyLlama/TinyLlama-1.1B20489.6
Polish models
sdadas/polish-gpt2-small204827.3
sdadas/polish-gpt2-medium204820.3
sdadas/polish-gpt2-large153618.0
sdadas/polish-gpt2-xl153616.6
Azurro/APT3-275M-Base204877.0
Azurro/APT3-500M-Base204850.5
Azurro/APT3-1B-Base204819.1
eryk-mazus/polka-1.1b20486.9
szymonrucinski/Curie-7B-v140964.8
OPI-PG/Qra-1b40966.1

Sample code

import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "eryk-mazus/polka-1.1b"

tokenizer = AutoTokenizer.from_pretrained(model_name, padding_side="left")
tokenizer.pad_token = tokenizer.eos_token

model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto", load_in_8bit=True)

prompt = """Przykładowe zapytanie do modelu"""

model_inputs = tokenizer([prompt], return_tensors="pt").to("cuda")
with torch.no_grad():
  generated_ids = model.generate(
    **model_inputs,
    max_new_tokens=512,
    do_sample=True,
    penalty_alpha=0.6,
    top_k=5
  )

output = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
print(output)
Downloads last month
45
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.
Invalid base_model specified in model card metadata. Needs to be a model id from hf.co/models.

Datasets used to train eryk-mazus/polka-1.1b

Space using eryk-mazus/polka-1.1b 1

Collection including eryk-mazus/polka-1.1b