--- license: gpl-3.0 datasets: - CohereForAI/aya_dataset language: - pl tags: - lobotomy --- **Polish-Lobotomy: An awful polish fine-tune** ============================================================ **Model Description** --------------- This fine-tuned Phi-3 model is the first attempt at a Polish fine-tune of Phi-3. It is very bad, probably because of the fine-tuning method (making the model learn a new language probably needs a full fine-tune) and the small dataset. - Ollama: [https://ollama.com/duckyblender/polish-lobotomy](https://ollama.com/duckyblender/polish-lobotomy) **Training Details** ----------------- * Trained on a single RTX 4060 for approximately 1 hour * Utilized 8-bit QLORA for efficient training * Despite the short training period, the model somehow managed to learn something (but not very well) ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6317acd6212fce5a3cd793f6/KnxTL_Ww3aYUrJz8kZ5Sz.jpeg) **Dataset** ------------ The model was trained on the Polish subset of the AYA dataset, which can be found at [https://huggingface.co/datasets/CohereForAI/aya_dataset](https://huggingface.co/datasets/CohereForAI/aya_dataset). **Prompt Template** ----------------- The prompt template used for this model is identical to the Phi 3 template. **Disclaimer** -------------- **Please be advised that this model's output may contain nonsensical responses. Viewer discretion is strongly advised (but not really necessary).** Use this model at your own risk, and please engage with the output responsibly (but let's be real, it's not like it's going to be useful for anything).