eachadea commited on
Commit
64205a5
1 Parent(s): 5ece6c7

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +90 -0
README.md ADDED
@@ -0,0 +1,90 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ tags:
5
+ - llama-2
6
+ - self-instruct
7
+ - distillation
8
+ - synthetic instruction
9
+ license:
10
+ - mit
11
+ ---
12
+
13
+ # Model Card: Nous-Hermes-Llama2-7b
14
+
15
+ Compute provided by our project sponsor Redmond AI, thank you! Follow RedmondAI on Twitter @RedmondAI.
16
+
17
+ ## Model Description
18
+
19
+ Nous-Hermes-Llama2-7b is a state-of-the-art language model fine-tuned on over 300,000 instructions. This model was fine-tuned by Nous Research, with Teknium leading the fine tuning process and dataset curation, Redmond AI sponsoring the compute, and several other contributors.
20
+
21
+ This Hermes model uses the exact same dataset as Hermes on Llama-1. This is to ensure consistency between the old Hermes and new, for anyone who wanted to keep Hermes as similar to the old one, just more capable.
22
+
23
+ This model stands out for its long responses, lower hallucination rate, and absence of OpenAI censorship mechanisms. The fine-tuning process was performed with a 4096 sequence length on an 8x a100 80GB DGX machine.
24
+
25
+
26
+ ## Model Training
27
+
28
+ The model was trained almost entirely on synthetic GPT-4 outputs. Curating high quality GPT-4 datasets enables incredibly high quality in knowledge, task completion, and style.
29
+
30
+ This includes data from diverse sources such as GPTeacher, the general, roleplay v1&2, code instruct datasets, Nous Instruct & PDACTL (unpublished), and several others, detailed further below
31
+
32
+ ## Collaborators
33
+ The model fine-tuning and the datasets were a collaboration of efforts and resources between Teknium, Karan4D, Emozilla, Huemin Art, and Redmond AI.
34
+
35
+ Special mention goes to @winglian for assisting in some of the training issues.
36
+
37
+ Huge shoutout and acknowledgement is deserved for all the dataset creators who generously share their datasets openly.
38
+
39
+ Among the contributors of datasets:
40
+ - GPTeacher was made available by Teknium
41
+ - Wizard LM by nlpxucan
42
+ - Nous Research Instruct Dataset was provided by Karan4D and HueminArt.
43
+ - GPT4-LLM and Unnatural Instructions were provided by Microsoft
44
+ - Airoboros dataset by jondurbin
45
+ - Camel-AI's domain expert datasets are from Camel-AI
46
+ - CodeAlpaca dataset by Sahil 2801.
47
+
48
+ If anyone was left out, please open a thread in the community tab.
49
+
50
+ ## Prompt Format
51
+
52
+ The model follows the Alpaca prompt format:
53
+ ```
54
+ ### Instruction:
55
+ <prompt>
56
+
57
+ ### Response:
58
+ <leave a newline blank for model to respond>
59
+
60
+ ```
61
+
62
+ or
63
+
64
+ ```
65
+ ### Instruction:
66
+ <prompt>
67
+
68
+ ### Input:
69
+ <additional context>
70
+
71
+ ### Response:
72
+ <leave a newline blank for model to respond>
73
+
74
+ ```
75
+
76
+ ## Benchmark Results
77
+ Coming soon
78
+
79
+ ## Resources for Applied Use Cases:
80
+ For an example of a back and forth chatbot using huggingface transformers and discord, check out: https://github.com/teknium1/alpaca-discord
81
+ For an example of a roleplaying discord chatbot, check out this: https://github.com/teknium1/alpaca-roleplay-discordbot
82
+
83
+ LM Studio is a good choice for a chat interface that supports GGML versions (to come)
84
+
85
+ ## Future Plans
86
+ We plan to continue to iterate on both more high quality data, and new data filtering techniques to eliminate lower quality data going forward.
87
+
88
+ ## Model Usage
89
+ The model is available for download on Hugging Face. It is suitable for a wide range of language tasks, from generating creative text to understanding and following complex instructions.
90
+