BramVanroy commited on
Commit
0dea749
•
1 Parent(s): 472bc9f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -10
README.md CHANGED
@@ -3,23 +3,34 @@ license: mit
3
  base_model: microsoft/phi-2
4
  tags:
5
  - trl
6
- - sft
7
- - generated_from_trainer
 
8
  datasets:
9
- - arrow
 
10
  model-index:
11
  - name: fietje-2b
12
  results: []
 
 
 
13
  ---
14
 
15
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
- should probably proofread and complete it, then remove this comment. -->
17
 
18
- # fietje-2b
 
 
19
 
20
- This model is a fine-tuned version of [microsoft/phi-2](https://huggingface.co/microsoft/phi-2) on the arrow dataset.
21
- It achieves the following results on the evaluation set:
22
- - Loss: 1.4013
 
 
 
 
 
 
23
 
24
  ## Model description
25
 
@@ -69,4 +80,4 @@ The following hyperparameters were used during training:
69
  - Transformers 4.39.1
70
  - Pytorch 2.1.2+cu121
71
  - Datasets 2.18.0
72
- - Tokenizers 0.15.2
 
3
  base_model: microsoft/phi-2
4
  tags:
5
  - trl
6
+ - conversational
7
+ - fietje
8
+ - alignment-handbook
9
  datasets:
10
+ - uonlp/CulturaX
11
+ - wikimedia/wikipedia
12
  model-index:
13
  - name: fietje-2b
14
  results: []
15
+ language:
16
+ - nl
17
+ pipeline_tag: text-generation
18
  ---
19
 
 
 
20
 
21
+ <p align="center" style="margin:0;padding:0">
22
+ <img src="https://huggingface.co/BramVanroy/fietje-2b/resolve/main/img/fietje-2b-banner.png" alt="Fietje banner" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
23
+ </p>
24
 
25
+ <div style="margin:auto; text-align:center">
26
+ <h1 style="margin-bottom: 0">Fietje 2B</h1>
27
+ <em>An open and efficient LLM for Dutch.</em>
28
+ </div>
29
+
30
+ > [!TIP]
31
+ > 🚀 Looking for the fast GGUF version? You can find it, and how to use it with `ollama`, [here](https://huggingface.co/BramVanroy/fietje-2b-GGUF). 🚀
32
+
33
+ This model is an adapted version of [microsoft/phi-2](https://huggingface.co/microsoft/phi-2), finetuned for Dutch text generation. It was continue-pretrained on 28B Dutch tokens, which includes the full Dutch component of Wikipedia and supplemented with Dutch tokens from CulturaX. A newer version of this dataset can be found [here](https://huggingface.co/datasets/BramVanroy/wikipedia_culturax_dutch), which also describes the filtering that took place.
34
 
35
  ## Model description
36
 
 
80
  - Transformers 4.39.1
81
  - Pytorch 2.1.2+cu121
82
  - Datasets 2.18.0
83
+ - Tokenizers 0.15.2