Suparious commited on
Commit
0a7e27a
1 Parent(s): 81edf60

update model card

Browse files
Files changed (1) hide show
  1. README.md +70 -0
README.md CHANGED
@@ -1,3 +1,73 @@
1
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ tags:
3
+ - finetuned
4
+ - quantized
5
+ - 4-bit
6
+ - AWQ
7
+ - transformers
8
+ - pytorch
9
+ - mistral
10
+ - instruct
11
+ - text-generation
12
+ - conversational
13
+ - license:apache-2.0
14
+ - autotrain_compatible
15
+ - endpoints_compatible
16
+ - text-generation-inference
17
+ - region:us
18
+ - finetune
19
+ - chatml
20
+ - DPO
21
+ - RLHF
22
+ - gpt4
23
+ - synthetic data
24
+ - distillation
25
+ model-index:
26
+ - name: Nous-Hermes-2-Mistral-7B-DPO
27
+ results: []
28
+ datasets:
29
+ - teknium/OpenHermes-2.5
30
  license: apache-2.0
31
+ language:
32
+ - en
33
+ quantized_by: Suparious
34
+ pipeline_tag: text-generation
35
+ model_creator: NousResearch
36
+ model_name: Nous Hermes 2 - Mistral 7B - DPO
37
+ inference: false
38
+ prompt_template: '<|im_start|>system
39
+
40
+ {system_message}<|im_end|>
41
+
42
+ <|im_start|>user
43
+
44
+ {prompt}<|im_end|>
45
+
46
+ <|im_start|>assistant
47
+
48
+ '
49
  ---
50
+ # Nous Hermes 2 - Mistral 7B - DPO
51
+
52
+ - Model creator: [NousResearch](https://huggingface.co/NousResearch)
53
+ - Original model: [WestLake 7B v2](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B)
54
+
55
+ The model prior to DPO was trained on 1,000,000 instructions/chats of GPT-4 quality or better, primarily synthetic data as well as other high quality datasets, available from the repository [teknium/OpenHermes-2.5](https://huggingface.co/datasets/teknium/OpenHermes-2.5).
56
+
57
+ ```plaintext
58
+ @misc{Nous-Hermes-2-Mistral-7B-DPO,
59
+ url={[https://huggingface.co/NousResearch/Nous-Hermes-2-Mistral-7B-DPO](https://huggingface.co/NousResearch/Nous-Hermes-2-Mistral-7B-DPO)},
60
+ title={Nous Hermes 2 Mistral 7B DPO},
61
+ author={"Teknium", "theemozilla", "karan4d", "huemin_art"}
62
+ }
63
+ ```
64
+
65
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/PDleZIZK3vE3ATfXRRySv.png)
66
+
67
+ ## Model Description
68
+
69
+ Nous Hermes 2 on Mistral 7B DPO is the new flagship 7B Hermes! This model was DPO'd from [Teknium/OpenHermes-2.5-Mistral-7B](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B) and has improved across the board on all benchmarks tested - AGIEval, BigBench Reasoning, GPT4All, and TruthfulQA.
70
+
71
+ The model prior to DPO was trained on 1,000,000 instructions/chats of GPT-4 quality or better, primarily synthetic data as well as other high quality datasets, available from the repository [teknium/OpenHermes-2.5](https://huggingface.co/datasets/teknium/OpenHermes-2.5).
72
+
73
+ ## Thank you to FluidStack for sponsoring compute for this model