--- base_model: unsloth/deepseek-r1-distill-qwen-7b-unsloth-bnb-4bit tags: - text-generation-inference - transformers - unsloth - qwen2 - trl license: apache-2.0 language: - en datasets: - Vezora/Tested-143k-Python-Alpaca - iamtarun/python_code_instructions_18k_alpaca - jtatman/python-code-dataset-500k - flytech/python-codes-25k - fxmeng/CodeFeedback-Python105K --- A finetuned model trained on **5 datasets** with a total of **875928 rows**. This model was an **experiment**, as I wanted to train a model with a lot of python code and see the results. # Uploaded model - **Developed by:** Vo1dAbyss - **License:** apache-2.0 - **Finetuned from model :** unsloth/deepseek-r1-distill-qwen-7b-unsloth-bnb-4bit This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [](https://github.com/unslothai/unsloth)