teknium commited on
Commit
6f7b716
1 Parent(s): f8eac33

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -6
README.md CHANGED
@@ -25,8 +25,16 @@ language:
25
  Collective Cognition v1 is a Mistral model fine-tuned using just 100 GPT-4 chats shared on Collective Cognition.
26
 
27
  ## Special Features:
28
- - **Quick Training**: This model was trained in just 3 minutes on a single 4090 with a qlora.
29
  - **Limited Data**: Despite its exceptional performance, it was trained on only ONE HUNDRED data points, all of which were gathered from Collective Cognition, a platform reminiscent of ShareGPT.
 
 
 
 
 
 
 
 
30
 
31
  ## Dataset:
32
 
@@ -42,6 +50,20 @@ You can download the datasets created by Collective Cognition here: https://hugg
42
 
43
  The model follows a LIMA approach, by minimizing the base model's original training as little as possible and giving a small but very high quality dataset to enhance it's performance and style.
44
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
  ## Benchmarks:
46
 
47
  Collective Cognition v1.0 TruthfulQA:
@@ -96,11 +118,6 @@ Average: 33.08%
96
  ```
97
 
98
 
99
- ## Acknowledgements:
100
-
101
- Special thanks to @a16z and all contributors to the Collective Cognition dataset for making the development of this model possible.
102
-
103
-
104
  ## Licensing:
105
 
106
  Apache 2.0
 
25
  Collective Cognition v1 is a Mistral model fine-tuned using just 100 GPT-4 chats shared on Collective Cognition.
26
 
27
  ## Special Features:
28
+ - **Quick Training**: This model was trained in just 3 minutes on a single 4090 with a qlora, and competes with 70B scale Llama-2 Models at TruthfulQA.
29
  - **Limited Data**: Despite its exceptional performance, it was trained on only ONE HUNDRED data points, all of which were gathered from Collective Cognition, a platform reminiscent of ShareGPT.
30
+ - **Extreme TruthfulQA Benchmark**: The collective cognition models are competing strongly with top 70B models on the TruthfulQA benchmark despite the small dataset and qlora training!
31
+
32
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/-pnifxPcMeeUONyE3efo3.png)
33
+
34
+ ## Acknowledgements:
35
+
36
+ Special thanks to @a16z and all contributors to the Collective Cognition dataset for making the development of this model possible.
37
+
38
 
39
  ## Dataset:
40
 
 
50
 
51
  The model follows a LIMA approach, by minimizing the base model's original training as little as possible and giving a small but very high quality dataset to enhance it's performance and style.
52
 
53
+ ## Usage:
54
+
55
+ Prompt Format:
56
+ ```
57
+ USER: <prompt>
58
+ ASSISTANT:
59
+ ```
60
+ OR
61
+ ```
62
+ <system message>
63
+ USER: <prompt>
64
+ ASSISTANT:
65
+ ```
66
+
67
  ## Benchmarks:
68
 
69
  Collective Cognition v1.0 TruthfulQA:
 
118
  ```
119
 
120
 
 
 
 
 
 
121
  ## Licensing:
122
 
123
  Apache 2.0