Update README.md
Browse files
README.md
CHANGED
@@ -36,18 +36,10 @@ language:
|
|
36 |
</p>
|
37 |
</blockquote>
|
38 |
|
|
|
|
|
|
|
39 |
|
40 |
-
This model is a fine-tuned version of [BramVanroy/fietje-2b-sft](https://huggingface.co/BramVanroy/fietje-2b-sft) on the BramVanroy/ultra_feedback_dutch_cleaned and the BramVanroy/orca_dpo_pairs_dutch_cleaned datasets.
|
41 |
-
It achieves the following results on the evaluation set:
|
42 |
-
- Loss: 0.2842
|
43 |
-
- Rewards/chosen: -1.1549
|
44 |
-
- Rewards/rejected: -3.6363
|
45 |
-
- Rewards/accuracies: 0.8867
|
46 |
-
- Rewards/margins: 2.4815
|
47 |
-
- Logps/rejected: -657.6813
|
48 |
-
- Logps/chosen: -451.3364
|
49 |
-
- Logits/rejected: -1.2868
|
50 |
-
- Logits/chosen: -1.3528
|
51 |
|
52 |
## Model description
|
53 |
|
|
|
36 |
</p>
|
37 |
</blockquote>
|
38 |
|
39 |
+
This is the chat version of Fietje, a DPO-tuned (aligned) variant of the instruct version. Fietje is an adapated version of [microsoft/phi-2](https://huggingface.co/microsoft/phi-2), tailored to Dutch text generation by training on 28B tokens. It is small and efficient with a size of 2.7 billion parameters while performing almost on par with more powerful Dutch LLMs of twice its size like [GEITje 7B Ultra](https://huggingface.co/BramVanroy/GEITje-7B-ultra).
|
40 |
+
|
41 |
+
A thorough description of the creation and evaluation of Fietje as well as usage examples are available in [this Github repository](https://github.com/BramVanroy/fietje).
|
42 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
43 |
|
44 |
## Model description
|
45 |
|