RJuro commited on
Commit
1fd7bc7
1 Parent(s): f3625f0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -13,9 +13,10 @@ A Danish finetune of Zephyr-7b-alpha 😀 The idea with this model (apart from p
13
 
14
  Try it here [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/RJuro/courses/blob/main/notebooks/Kanelsnegl_v0_2_usecases.ipynb)
15
 
16
- <!--
17
- <img src="https://huggingface.co/RJuro/kanelsnegl-v0.1/resolve/main/kanelsnegl_banner.png" alt="Kanelsnegl Logo" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
18
- -->
 
19
  ## Model Description
20
  Base model: [Zephyr-7b-alpha](https://huggingface.co/HuggingFaceH4/zephyr-7b-alpha) finetuned on [DDSC/partial-danish-gigaword-no-twitter](https://huggingface.co/datasets/DDSC/partial-danish-gigaword-no-twitter). The training involved a maximum length of 512. QLora completion finetuning of all linear layers was also implemented. This model is mostly fun tinkering for personal learning purpose.
21
  This version got 4 times more fine-tuning than v0.1 [RJuro/kanelsnegl-v0.1](https://huggingface.co/RJuro/kanelsnegl-v0.1). It produces better Danish and follows complex prompts and instructions.
 
13
 
14
  Try it here [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/RJuro/courses/blob/main/notebooks/Kanelsnegl_v0_2_usecases.ipynb)
15
 
16
+
17
+ <img src="https://huggingface.co/RJuro/kanelsnegl-v0.2/resolve/main/banner_ks_s.jpg" alt="Kanelsnegl Logo" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
18
+
19
+
20
  ## Model Description
21
  Base model: [Zephyr-7b-alpha](https://huggingface.co/HuggingFaceH4/zephyr-7b-alpha) finetuned on [DDSC/partial-danish-gigaword-no-twitter](https://huggingface.co/datasets/DDSC/partial-danish-gigaword-no-twitter). The training involved a maximum length of 512. QLora completion finetuning of all linear layers was also implemented. This model is mostly fun tinkering for personal learning purpose.
22
  This version got 4 times more fine-tuning than v0.1 [RJuro/kanelsnegl-v0.1](https://huggingface.co/RJuro/kanelsnegl-v0.1). It produces better Danish and follows complex prompts and instructions.