--- datasets: - cognitivecomputations/dolphin - jondurbin/airoboros-2.2.1 - cognitivecomputations/dolphin-coder - teknium/openhermes - ise-uiuc/Magicoder-OSS-Instruct-75K - ise-uiuc/Magicoder-Evol-Instruct-110K - m-a-p/Code-Feedback - m-a-p/CodeFeedback-Filtered-Instruction - microsoft/orca-math-word-problems-200k language: - en license: bigcode-openrail-m --- DolphinCoder StarCoder2 7b 🐬 sponsored by [latitude.sh](https://www.latitude.sh/). Join our Discord! https://discord.gg/cognitivecomputations This model is based on StarCoder2-7b and is subject to bigcode-openrail-m license. This Dolphin is *really good* at coding, I trained with a lot of coding data. This model is uncensored. I have filtered the dataset to remove alignment and bias. This makes the model more compliant. You are advised to implement your own alignment layer before exposing the model as a service. It will be highly compliant to any requests, even unethical ones. Please read my blog post about uncensored models. https://erichartford.com/uncensored-models You are responsible for any content you create using this model. Enjoy responsibly. ## Training It took 2 days to train 3 epochs on 8x L40S's using qLoRA and Axolotl Prompt format: This model uses ChatML prompt format. ``` <|im_start|>system You are DolphinCoder, a helpful AI programming assistant.<|im_end|> <|im_start|>user {prompt}<|im_end|> <|im_start|>assistant ``` Example: ``` <|im_start|>system You are DolphinCoder, a master at software engineering and coding in any programming language. <|im_start|>user Please write me a program in golang that parses all the lines in a file, and reverses them character-wise, and saves it to a new file. <|im_start|>assistant ``` ## Quantized models - [dagbs/-GGUF](https://huggingface.co/dagbs/dolphincoder-starcoder2-7b-GGUF) ## Gratitude - This model was made possible by the generous sponsorship of [latitude.sh](https://www.latitude.sh/). - Welcome Microsoft to Open Source AI! Thank you for the Orca-Math Dataset! - Huge thank you to [BigCode](https://www.bigcode-project.org/) for training and publishing the weights of StarCoder2 - HUGE Thank you to the dataset authors: @ise-uiuc, @teknium, @m-a-p - And HUGE thanks to @winglian and the Axolotl contributors for making the best training framework! - [Built with Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) - Thank you to all the other people in the Open Source AI community who have taught me and helped me along the way. ## Example Output [If you would like to financially support my efforts](https://ko-fi.com/erichartford) [swag](https://fa7113.myshopify.com/)