BramVanroy commited on
Commit
529e879
·
verified ·
1 Parent(s): 677da9d

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +44 -0
README.md ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ language:
4
+ - nl
5
+ tags:
6
+ - gguf
7
+ - llamacpp
8
+ - dpo
9
+ - geitje
10
+ - conversational
11
+ datasets:
12
+ - BramVanroy/ultra_feedback_dutch
13
+ ---
14
+
15
+ <img src="https://huggingface.co/BramVanroy/GEITje-7B-ultra/resolve/main/geitje-ultra-banner.png" alt="GEITje Ultra banner" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
16
+
17
+
18
+ # GEITje 7B ultra (GGUF version)
19
+
20
+ This is a `Q5_K_M` GGUF version of [BramVanroy/GEITje-7B-ultra](https://huggingface.co/BramVanroy/GEITje-7B-ultra). For more information on the model, data, licensing, usage, see the main model's README.
21
+
22
+ ## Repro
23
+
24
+ Assuming you have installed and build llama cpp, current working directory is the `build` directory in llamacpp.
25
+
26
+ Download initial model (probaby a huggingface-cli alternative exists, too...)
27
+
28
+ ```python
29
+ from huggingface_hub import snapshot_download
30
+ model_id = "BramVanroy/GEITje-7B-ultra"
31
+ snapshot_download(repo_id=model_id, local_dir="geitje-ultra-hf", local_dir_use_symlinks=False)
32
+ ```
33
+
34
+ Convert to GGML format
35
+
36
+ ```shell
37
+ # Convert to GGML format
38
+ python convert.py build/geitje-ultra-hf/
39
+
40
+ cd build
41
+
42
+ # Quantize to Q5_K_M
43
+ bin/quantize geitje-ultra-hf/ggml-model-f32.gguf geitje-ultra-hf/GEITje-7B-ultra-Q5_K_M.gguf Q5_K_M
44
+ ```