Fizzarolli's picture
Update README.md
6ef240c verified
|
raw
history blame
1.67 kB
metadata
base_model:
  - Qwen/Qwen2.5-72B
tags:
  - roleplay
  - storywriting
  - qwen2.5
  - finetune
  - transformers
  - pytorch

Zeus Labs ~ Chronos-Platinum-72B

image/webp

Qwen 2 72B base model, trained for two epochs on the Chronos Gold finetuning dataset.

Instruct Template

This model uses ChatML - below is an example. It is a preset in many frontends.

<|im_start|>system
You are a helpful assistant<|im_end|>
<|im_start|>user
Hello there!<|im_end|>
<|im_start|>assistant
Hi! I'm an AI assistant, designed to help people like you with all sorts of tasks. Is there anything you need help with?<|im_end|>
<|im_start|>user
I was wondering how transformers work?<|im_end|>
<|im_start|>assistant

Quantization

Please note that we tested this model with a 4.5bpw EXL2 quant. Results are not expected to be the same when going below this quanitzation.

LlamaCPP

TODO!

Exllama2

TODO!

FP8

TODO!

Sampling Settings

Here are some settings that work well with this model:

TODO!

Credit

Thank you to my team consisting of @Fizzarolli, @ToastyPigeon and myself @elinas.

Additional thanks to @AlpinDale and the rest of the PygmalionAI team for graciously providing the compute to finetune this model!

Additional Details

If you have any questions or concerns, please post in the community tab.

DISCLAIMER: Outputs generated by the model are not reflective of our views.