File size: 1,315 Bytes
57116f8
 
 
 
 
 
 
 
 
 
 
 
 
 
bccfc09
57116f8
 
 
179fb61
 
 
 
57116f8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
---
base_model:
- nbeerbower/mistral-nemo-bophades-12B
- nbeerbower/mistral-nemo-gutenberg-12B-v3
license: apache-2.0
library_name: transformers
tags:
- merge
- roleplay
- not-for-all-audiences
---

# Magnum-Instruct-DPO-12B

Similar 50/50 merge like the other Magnum-Instruct, but using model variants that have had extra dpo/orpo training on top of them beforehand. Can't say if it's better or not comparatively speaking to just using the original models yet, but it seamed fine enough during my limited testing and worth the upload for now as an alternative.

Big thanks to the MistralAI and Anthracite/SillyTilly teams for the original models used, plus nbeerbower for the extra training done as well!

GGUF quants provided by mradermacher:

https://huggingface.co/mradermacher/Magnum-Instruct-DPO-12B-GGUF

## Settings

Temperature @ 0.7

Min-P @ 0.02

Smoothing Factor @ 0.3

Smoothing Curve @ 1.5

DRY Multiplier (plus standard DRY settings) @ 0.8

Skip Special Tokens @ On

Everything else @ Off

### Prompt Format: Nemo-Mistral

```
[INST] user prompt[/INST] character response</s>[INST] user prompt[/INST]
```

### Models Merged

The following models were included in the merge:

https://huggingface.co/nbeerbower/mistral-nemo-bophades-12B

https://huggingface.co/nbeerbower/mistral-nemo-gutenberg-12B-v3