Update README.md
Browse files
README.md
CHANGED
@@ -26,15 +26,17 @@ I did not except this repo to blow up and now all the training scripts depend on
|
|
26 |
|
27 |
Now for the magic trained finetune that runs at insane speeds:
|
28 |
|
29 |
-
```verilog
|
30 |
-
wget https://huggingface.co/nisten/Biggie-SmoLlm-0.15B-Base/resolve/main/biggie_groked_int8_q8_0.gguf
|
31 |
-
```
|
32 |
The settings are very finicky so be careful with your experimentation
|
33 |
-
```
|
34 |
-
./llama-cli -fa -b 512 -ctv q8_0 -ctk q8_0 --min-p 0.3 --top-p 0.85 --keep -1
|
|
|
|
|
|
|
|
|
35 |
```
|
|
|
36 |
|
37 |
-
|
38 |
Model is more coherent.
|
39 |
|
40 |
The temperature settings and min p etc need to be adjusted but even at default temp0 it was coherent for first 100 tokens.
|
|
|
26 |
|
27 |
Now for the magic trained finetune that runs at insane speeds:
|
28 |
|
|
|
|
|
|
|
29 |
The settings are very finicky so be careful with your experimentation
|
30 |
+
```verilog
|
31 |
+
./llama-cli -fa -b 512 -ctv q8_0 -ctk q8_0 --min-p 0.3 --top-p 0.85 --keep -1 \
|
32 |
+
-p "You are a NASA JPL Scientists. Human: I want to bring my cat to mars." \
|
33 |
+
--in-prefix "<|im_start|>Human:" --reverse-prompt "Human:" \
|
34 |
+
-m biggie_groked_int8_q8_0.gguf -co -cnv \
|
35 |
+
-c 1024 -n 700 --temp 1.5 -ngl 0 -t 1
|
36 |
```
|
37 |
+
Yup, that's no gpu, 1 cpu core.
|
38 |
|
39 |
+
This base model was built one via semi-automated continuous merging to figure out the recipe.
|
40 |
Model is more coherent.
|
41 |
|
42 |
The temperature settings and min p etc need to be adjusted but even at default temp0 it was coherent for first 100 tokens.
|