File size: 685 Bytes
1f3d4a4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
# superhot-13b-8k-no-rlhf-test-GGML

Merged base LLaMA and LoRA with this:
https://github.com/tloen/alpaca-lora

Base LLaMA 13B:
https://huggingface.co/huggyllama/llama-13b

SuperHOT 13B 8k no-rlhf-test LoRA:
https://huggingface.co/kaiokendev/superhot-13b-8k-no-rlhf-test

``` sh
BASE_MODEL=huggyllama_llama-13b LORA=kaiokendev_superhot-13b-8k-no-rlhf-test python export_hf_checkpoint.py
```

Converted and quantized with llama.cpp commit `447ccbe`:

``` sh
python convert.py superhot-13b-8k-safetensors --outtype f32 --outfile superhot-13b-8k-no-rlhf-test.ggmlv3.f32.bin
./bin/quantize superhot-13b-8k-no-rlhf-test.ggmlv3.f32.bin superhot-13b-8k-no-rlhf-test.ggmlv3.Q2_K.bin Q2_K
```