BramVanroy commited on
Commit
ae75069
·
verified ·
1 Parent(s): 7f1cfac

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -11
README.md CHANGED
@@ -36,18 +36,10 @@ language:
36
  </p>
37
  </blockquote>
38
 
 
 
 
39
 
40
- This model is a fine-tuned version of [BramVanroy/fietje-2b-sft](https://huggingface.co/BramVanroy/fietje-2b-sft) on the BramVanroy/ultra_feedback_dutch_cleaned and the BramVanroy/orca_dpo_pairs_dutch_cleaned datasets.
41
- It achieves the following results on the evaluation set:
42
- - Loss: 0.2842
43
- - Rewards/chosen: -1.1549
44
- - Rewards/rejected: -3.6363
45
- - Rewards/accuracies: 0.8867
46
- - Rewards/margins: 2.4815
47
- - Logps/rejected: -657.6813
48
- - Logps/chosen: -451.3364
49
- - Logits/rejected: -1.2868
50
- - Logits/chosen: -1.3528
51
 
52
  ## Model description
53
 
 
36
  </p>
37
  </blockquote>
38
 
39
+ This is the chat version of Fietje, a DPO-tuned (aligned) variant of the instruct version. Fietje is an adapated version of [microsoft/phi-2](https://huggingface.co/microsoft/phi-2), tailored to Dutch text generation by training on 28B tokens. It is small and efficient with a size of 2.7 billion parameters while performing almost on par with more powerful Dutch LLMs of twice its size like [GEITje 7B Ultra](https://huggingface.co/BramVanroy/GEITje-7B-ultra).
40
+
41
+ A thorough description of the creation and evaluation of Fietje as well as usage examples are available in [this Github repository](https://github.com/BramVanroy/fietje).
42
 
 
 
 
 
 
 
 
 
 
 
 
43
 
44
  ## Model description
45