--- base_model: mistralai/Mistral-Nemo-Base-2407 language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - mistral - trl - sft datasets: - mpasila/LimaRP-PIPPA-Mix-8K-Context - grimulkan/LimaRP-augmented - KaraKaraWitch/PIPPA-ShareGPT-formatted --- LoRA trained in 4-bit with 8k context using [mistralai/Mistral-Nemo-Base-2407](https://huggingface.co/mistralai/Mistral-Nemo-Base-2407/) as the base model for 1 epoch. Dataset used is [mpasila/LimaRP-PIPPA-Mix-8K-Context](https://huggingface.co/datasets/mpasila/LimaRP-PIPPA-Mix-8K-Context) which was made using [grimulkan/LimaRP-augmented](https://huggingface.co/datasets/grimulkan/LimaRP-augmented) and [KaraKaraWitch/PIPPA-ShareGPT-formatted](https://huggingface.co/datasets/KaraKaraWitch/PIPPA-ShareGPT-formatted). Merged from this LoRA: [mpasila/Mistral-LiPPA-LoRA-12B](https://huggingface.co/mpasila/Mistral-LiPPA-LoRA-12B) So uhh it does kinda work, maybe not the best datasets but uhh it's something. ### Prompt format: Llama 3 Instruct Unsloth changed assistant to gpt and user to human. # Uploaded model - **Developed by:** mpasila - **License:** apache-2.0 - **Finetuned from model :** unsloth/mistral-nemo-base-2407-bnb-4bit This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [](https://github.com/unslothai/unsloth)