--- base_model: - Qwen/Qwen2.5-72B tags: - roleplay - storywriting - qwen2.5 - finetune - transformers - pytorch --- # Zeus Labs ~ Chronos-Platinum-72B ![image/webp](https://cdn-uploads.huggingface.co/production/uploads/634262af8d8089ebaefd410e/tR2GBVNmTEWt8HD6LQF3U.webp) Qwen 2 72B base model, trained for two epochs on the Chronos Gold finetuning dataset. ## Instruct Template This model uses `ChatML` - below is an example. It is a preset in many frontends. ``` <|im_start|>system You are a helpful assistant<|im_end|> <|im_start|>user Hello there!<|im_end|> <|im_start|>assistant Hi! I'm an AI assistant, designed to help people like you with all sorts of tasks. Is there anything you need help with?<|im_end|> <|im_start|>user I was wondering how transformers work?<|im_end|> <|im_start|>assistant ``` ## Quantization Please note that we tested this model with a 4.5bpw EXL2 quant. Results are not expected to be the same when going below this quanitzation. #### LlamaCPP TODO! #### Exllama2 TODO! #### FP8 TODO! ## Sampling Settings Here are some settings that work well with this model: ``` TODO! ``` ## Credit Thank you to my team consisting of [@Fizzarolli](https://huggingface.co/Fizzarolli), [@ToastyPigeon](https://huggingface.co/ToastyPigeon) and myself [@elinas](https://huggingface.co/elinas). Additional thanks to [@AlpinDale](https://huggingface.co/AlpinDale) and the rest of the PygmalionAI team for graciously providing the compute to finetune this model! ## Additional DetailsĀ  If you have any questions or concerns, please post in the community tab. DISCLAIMER: Outputs generated by the model are not reflective of our views.