teknium commited on
Commit
fe31207
1 Parent(s): 61fa08e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: upstage/SOLAR-10.7B-v1.0
3
  tags:
4
  - Mixtral
5
  - instruct
@@ -18,7 +18,7 @@ language:
18
  - en
19
  ---
20
 
21
- # Nous Hermes 2 - Solar 10.7B
22
 
23
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/qVRnEDL_BUEWulUvWBD95.png)
24
 
@@ -28,6 +28,8 @@ Nous Hermes 2 Mixtral 7bx8 DPO is the new flagship Nous Research model trained o
28
 
29
  The model was trained on over 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks.
30
 
 
 
31
  # Table of Contents
32
  1. [Example Outputs](#example-outputs)
33
  2. [Benchmark Results](#benchmark-results)
 
1
  ---
2
+ base_model: mistralai/Mixtral-8x7B-v0.1
3
  tags:
4
  - Mixtral
5
  - instruct
 
18
  - en
19
  ---
20
 
21
+ # Nous Hermes 2 - Mixtral 8x7B-DPO
22
 
23
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/qVRnEDL_BUEWulUvWBD95.png)
24
 
 
28
 
29
  The model was trained on over 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks.
30
 
31
+ This is the SFT + DPO version of Mixtral Hermes 2, we will also be providing an SFT only version, for people to find which works best for them.
32
+
33
  # Table of Contents
34
  1. [Example Outputs](#example-outputs)
35
  2. [Benchmark Results](#benchmark-results)