matthieumeeus97 commited on
Commit
308ca56
·
verified ·
1 Parent(s): 9b8c17f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -43,7 +43,7 @@ We provide 6 variants (of which 3 base and 3 instruction-tuned models):
43
  - **Llama-3-ChocoLlama-8B-base** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-base)): A language-adapted version of Meta's Llama-8-8B, fine-tuned on the same Dutch dataset as ChocoLlama-2-7B-base, again using LoRa.
44
  - **Llama-3-ChocoLlama-instruct** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-instruct)): An instruction-tuned version of Llama-3-ChocoLlama-8B-base, fine-tuned on the same dataset as ChocoLlama-2-7B-instruct, again using SFT followed by DPO.
45
 
46
- For benchmark results for all models, including compared to their base models and other Dutch LLMs, we refer to our paper [here](some_url).
47
 
48
  ### Model Description
49
 
@@ -55,8 +55,8 @@ For benchmark results for all models, including compared to their base models an
55
 
56
  ### Model Sources
57
 
58
- - **Repository:** Will be released soon.
59
- - **Paper:** Will be released soon.
60
 
61
  ## Uses
62
 
 
43
  - **Llama-3-ChocoLlama-8B-base** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-base)): A language-adapted version of Meta's Llama-8-8B, fine-tuned on the same Dutch dataset as ChocoLlama-2-7B-base, again using LoRa.
44
  - **Llama-3-ChocoLlama-instruct** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-instruct)): An instruction-tuned version of Llama-3-ChocoLlama-8B-base, fine-tuned on the same dataset as ChocoLlama-2-7B-instruct, again using SFT followed by DPO.
45
 
46
+ For benchmark results for all models, including compared to their base models and other Dutch LLMs, we refer to our paper [here](https://arxiv.org/pdf/2412.07633).
47
 
48
  ### Model Description
49
 
 
55
 
56
  ### Model Sources
57
 
58
+ - **Repository:** [on Github here](https://github.com/ChocoLlamaModel/ChocoLlama).
59
+ - **Paper:** [on ArXiv here](https://arxiv.org/pdf/2412.07633).
60
 
61
  ## Uses
62