Update README.md
Browse files
README.md
CHANGED
@@ -25,8 +25,16 @@ language:
|
|
25 |
Collective Cognition v1 is a Mistral model fine-tuned using just 100 GPT-4 chats shared on Collective Cognition.
|
26 |
|
27 |
## Special Features:
|
28 |
-
- **Quick Training**: This model was trained in just 3 minutes on a single 4090 with a qlora.
|
29 |
- **Limited Data**: Despite its exceptional performance, it was trained on only ONE HUNDRED data points, all of which were gathered from Collective Cognition, a platform reminiscent of ShareGPT.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
30 |
|
31 |
## Dataset:
|
32 |
|
@@ -42,6 +50,20 @@ You can download the datasets created by Collective Cognition here: https://hugg
|
|
42 |
|
43 |
The model follows a LIMA approach, by minimizing the base model's original training as little as possible and giving a small but very high quality dataset to enhance it's performance and style.
|
44 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
45 |
## Benchmarks:
|
46 |
|
47 |
Collective Cognition v1.0 TruthfulQA:
|
@@ -96,11 +118,6 @@ Average: 33.08%
|
|
96 |
```
|
97 |
|
98 |
|
99 |
-
## Acknowledgements:
|
100 |
-
|
101 |
-
Special thanks to @a16z and all contributors to the Collective Cognition dataset for making the development of this model possible.
|
102 |
-
|
103 |
-
|
104 |
## Licensing:
|
105 |
|
106 |
Apache 2.0
|
|
|
25 |
Collective Cognition v1 is a Mistral model fine-tuned using just 100 GPT-4 chats shared on Collective Cognition.
|
26 |
|
27 |
## Special Features:
|
28 |
+
- **Quick Training**: This model was trained in just 3 minutes on a single 4090 with a qlora, and competes with 70B scale Llama-2 Models at TruthfulQA.
|
29 |
- **Limited Data**: Despite its exceptional performance, it was trained on only ONE HUNDRED data points, all of which were gathered from Collective Cognition, a platform reminiscent of ShareGPT.
|
30 |
+
- **Extreme TruthfulQA Benchmark**: The collective cognition models are competing strongly with top 70B models on the TruthfulQA benchmark despite the small dataset and qlora training!
|
31 |
+
|
32 |
+
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/-pnifxPcMeeUONyE3efo3.png)
|
33 |
+
|
34 |
+
## Acknowledgements:
|
35 |
+
|
36 |
+
Special thanks to @a16z and all contributors to the Collective Cognition dataset for making the development of this model possible.
|
37 |
+
|
38 |
|
39 |
## Dataset:
|
40 |
|
|
|
50 |
|
51 |
The model follows a LIMA approach, by minimizing the base model's original training as little as possible and giving a small but very high quality dataset to enhance it's performance and style.
|
52 |
|
53 |
+
## Usage:
|
54 |
+
|
55 |
+
Prompt Format:
|
56 |
+
```
|
57 |
+
USER: <prompt>
|
58 |
+
ASSISTANT:
|
59 |
+
```
|
60 |
+
OR
|
61 |
+
```
|
62 |
+
<system message>
|
63 |
+
USER: <prompt>
|
64 |
+
ASSISTANT:
|
65 |
+
```
|
66 |
+
|
67 |
## Benchmarks:
|
68 |
|
69 |
Collective Cognition v1.0 TruthfulQA:
|
|
|
118 |
```
|
119 |
|
120 |
|
|
|
|
|
|
|
|
|
|
|
121 |
## Licensing:
|
122 |
|
123 |
Apache 2.0
|