nisten commited on
Commit
6cab427
·
verified ·
1 Parent(s): 24116da

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -6
README.md CHANGED
@@ -26,15 +26,17 @@ I did not except this repo to blow up and now all the training scripts depend on
26
 
27
  Now for the magic trained finetune that runs at insane speeds:
28
 
29
- ```verilog
30
- wget https://huggingface.co/nisten/Biggie-SmoLlm-0.15B-Base/resolve/main/biggie_groked_int8_q8_0.gguf
31
- ```
32
  The settings are very finicky so be careful with your experimentation
33
- ```bash
34
- ./llama-cli -fa -b 512 -ctv q8_0 -ctk q8_0 --min-p 0.3 --top-p 0.85 --keep -1 -p "You are a NASA JPL Scientists. Human: I want to bring my cat to mars." -m biggie_groked_int8_q8_0.gguf -co -cnv --in-prefix "<|im_start|>Human:" --reverse-prompt "Human:" -c 1024 -n 700 --temp 1.5 -ngl 0 -t 1
 
 
 
 
35
  ```
 
36
 
37
- Done via semi-automated continuous merging to figure out the recipe.
38
  Model is more coherent.
39
 
40
  The temperature settings and min p etc need to be adjusted but even at default temp0 it was coherent for first 100 tokens.
 
26
 
27
  Now for the magic trained finetune that runs at insane speeds:
28
 
 
 
 
29
  The settings are very finicky so be careful with your experimentation
30
+ ```verilog
31
+ ./llama-cli -fa -b 512 -ctv q8_0 -ctk q8_0 --min-p 0.3 --top-p 0.85 --keep -1 \
32
+ -p "You are a NASA JPL Scientists. Human: I want to bring my cat to mars." \
33
+ --in-prefix "<|im_start|>Human:" --reverse-prompt "Human:" \
34
+ -m biggie_groked_int8_q8_0.gguf -co -cnv \
35
+ -c 1024 -n 700 --temp 1.5 -ngl 0 -t 1
36
  ```
37
+ Yup, that's no gpu, 1 cpu core.
38
 
39
+ This base model was built one via semi-automated continuous merging to figure out the recipe.
40
  Model is more coherent.
41
 
42
  The temperature settings and min p etc need to be adjusted but even at default temp0 it was coherent for first 100 tokens.