mpasila's picture
Update README.md
0b97593 verified
|
raw
history blame contribute delete
No virus
935 Bytes
metadata
base_model: mistralai/Mistral-Nemo-Base-2407
license: apache-2.0
datasets:
  - BeaverAI/Nemo-Inst-Tune-ds
language:
  - en
library_name: transformers

This is an ExLlamaV2 quantized model in 3.5bpw of BeaverAI/mistral-dory-12b using the default calibration dataset with 8192 context length.

Original Model card:

Dory 12b

redone instruct finetune of mistral nemo 12b. not (E)RP-focused, leave that to drummer.

image/gif

thanks to twisted for the compute :3

Prompting

alpaca-like:

### System:
[Optional system prompt]

### Instruction:
[Query]

### Response:
[Response]<EOT>

### Instruction:
[...]

Training details

Rank 64 QDoRA, trained on primarily Claude and Gemma 2 multiturn data (it's midnight and I'll probably write more details tomorrow)