SanjiWatsuki commited on
Commit
b4268a4
1 Parent(s): 8e46112

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +65 -0
README.md ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ ---
4
+
5
+ ![image/png](https://huggingface.co/SanjiWatsuki/Kunoichi-7B/resolve/main/assets/kunoichi.png)
6
+
7
+ <!-- description start -->
8
+ ## Description
9
+
10
+ This repository hosts **Kunoichi-DPO-7B**, a DPO finetune using Intel's Orca pairs with the Alpaca template on Kunoichi-7B. This model is targeted at general use. In my testing, it has stronger reasoning and instruction following capabilities than Kunoichi-7B but it may be worse for roleplaying purposes due to the alignment from the Orca dataset.
11
+
12
+ This model is undergoing benchmark testing and I will update the model page with the finalized results.
13
+
14
+ | Model | MT Bench | EQ Bench | MMLU | Logic Test |
15
+ |----------------------|----------|----------|---------|-------------|
16
+ | GPT-4-Turbo | 9.32 | - | - | - |
17
+ | GPT-4 | 8.99 | 62.52 | 86.4 | 0.86 |
18
+ | **Kunoichi-DPO-7B** | **8.29** | **41.60** | - | **0.59** |
19
+ | **Kunoichi-7B** | **8.14** | **44.32** | **64.9** | **0.58** |
20
+ | Starling-7B | 8.09 | - | 63.9 | 0.51 |
21
+ | Claude-2 | 8.06 | 52.14 | 78.5 | - |
22
+ | Silicon-Maid-7B | 7.96 | 40.44 | 64.7 | 0.54 |
23
+ | Loyal-Macaroni-Maid-7B | 7.95 | 38.66 | 64.9 | 0.57 |
24
+ | GPT-3.5-Turbo | 7.94 | 50.28 | 70 | 0.57 |
25
+ | Claude-1 | 7.9 | - | 77 | - |
26
+ | Openchat-3.5 | 7.81 | 37.08 | 64.3 | 0.39 |
27
+ | Dolphin-2.6-DPO | 7.74 | 42.88 | 61.9 | 0.53 |
28
+ | Zephyr-7B-beta | 7.34 | 38.71 | 61.4 | 0.30 |
29
+ | Llama-2-70b-chat-hf | 6.86 | 51.56 | 63 | - |
30
+ | Neural-chat-7b-v3-1 | 6.84 | 43.61 | 62.4 | 0.30 |
31
+
32
+ | Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
33
+ |---|---:|---:|---:|---:|---:|
34
+ | [**Kunoichi-DPO-7B**]|-| -| -| -| -|
35
+ | [Kunoichi-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-7B)|57.54| 44.99| 74.86| 63.72| 46.58|
36
+ | [OpenPipe/mistral-ft-optimized-1218](https://huggingface.co/OpenPipe/mistral-ft-optimized-1218)| 56.85 | 44.74 | 75.6 | 59.89 | 47.17 |
37
+ | [Silicon-Maid-7B](https://huggingface.co/SanjiWatsuki/Silicon-Maid-7B) | 56.45| 44.74| 74.26| 61.5| 45.32|
38
+ | [mlabonne/NeuralHermes-2.5-Mistral-7B](https://huggingface.co/mlabonne/NeuralHermes-2.5-Mistral-7B) | 53.51 | 43.67 | 73.24 | 55.37 | 41.76 |
39
+ | [teknium/OpenHermes-2.5-Mistral-7B](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B) | 52.42 | 42.75 | 72.99 | 52.99 | 40.94 |
40
+ | [openchat/openchat_3.5](https://huggingface.co/openchat/openchat_3.5) | 51.34 | 42.67 | 72.92 | 47.27 | 42.51 |
41
+ | [berkeley-nest/Starling-LM-7B-alpha](https://huggingface.co/berkeley-nest/Starling-LM-7B-alpha) | 51.16 | 42.06 | 72.72 | 47.33 | 42.53 |
42
+ | [HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) | 50.99 | 37.33 | 71.83 | 55.1 | 39.7 |
43
+
44
+ The model is intended to be used with up to an 8k context window. Using a NTK RoPE alpha of 2.6, the model can be used experimentally up to a 16k context window.
45
+
46
+ <!-- description end -->
47
+ <!-- prompt-template start -->
48
+ ## Prompt template: Custom format, or Alpaca
49
+
50
+ ### Alpaca:
51
+ ```
52
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
53
+
54
+ ### Instruction:
55
+ {prompt}
56
+
57
+ ### Response:
58
+ ```
59
+
60
+ ### SillyTavern format:
61
+ I found the best SillyTavern results from using the Noromaid template.
62
+
63
+ SillyTavern config files: [Context](https://files.catbox.moe/ifmhai.json), [Instruct](https://files.catbox.moe/ttw1l9.json).
64
+
65
+ Additionally, here is my highly recommended [Text Completion preset](https://huggingface.co/SanjiWatsuki/Loyal-Macaroni-Maid-7B/blob/main/Characters/MinP.json). You can tweak this by adjusting temperature up or dropping min p to boost creativity or raise min p to increase stability. You shouldn't need to touch anything else!