polish-lobotomy / README.md
DuckyBlender's picture
Create README.md
1fe49dc verified
---
license: gpl-3.0
datasets:
- CohereForAI/aya_dataset
language:
- pl
tags:
- lobotomy
---
**Polish-Lobotomy: An awful polish fine-tune**
============================================================
**Model Description**
---------------
This fine-tuned Phi-3 model is the first attempt at a Polish fine-tune of Phi-3. It is very bad, probably because of the fine-tuning method (making the model learn a new language probably needs a full fine-tune) and the small dataset.
- Ollama: [https://ollama.com/duckyblender/polish-lobotomy](https://ollama.com/duckyblender/polish-lobotomy)
**Training Details**
-----------------
* Trained on a single RTX 4060 for approximately 1 hour
* Utilized 8-bit QLORA for efficient training
* Despite the short training period, the model somehow managed to learn something (but not very well)
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6317acd6212fce5a3cd793f6/KnxTL_Ww3aYUrJz8kZ5Sz.jpeg)
**Dataset**
------------
The model was trained on the Polish subset of the AYA dataset, which can be found at [https://huggingface.co/datasets/CohereForAI/aya_dataset](https://huggingface.co/datasets/CohereForAI/aya_dataset).
**Prompt Template**
-----------------
The prompt template used for this model is identical to the Phi 3 template.
**Disclaimer**
--------------
**Please be advised that this model's output may contain nonsensical responses. Viewer discretion is strongly advised (but not really necessary).**
Use this model at your own risk, and please engage with the output responsibly (but let's be real, it's not like it's going to be useful for anything).