Update README.md
Browse files
README.md
CHANGED
@@ -21,7 +21,7 @@ license: other
|
|
21 |
|
22 |
These files are GPTQ 4bit model files for [Eric Hartford's WizardLM 13B V1.0 Uncensored](https://huggingface.co/ehartford/WizardLM-13B-V1.0-Uncensored) merged with [Kaio Ken's SuperHOT 8K](https://huggingface.co/kaiokendev/superhot-13b-8k-no-rlhf-test).
|
23 |
|
24 |
-
[Kaio Ken's SuperHOT
|
25 |
|
26 |
Note that `config.json` has been set to a sequence length of 8192. This can be modified to 4096 if you want to try with a smaller sequence length.
|
27 |
|
|
|
21 |
|
22 |
These files are GPTQ 4bit model files for [Eric Hartford's WizardLM 13B V1.0 Uncensored](https://huggingface.co/ehartford/WizardLM-13B-V1.0-Uncensored) merged with [Kaio Ken's SuperHOT 8K](https://huggingface.co/kaiokendev/superhot-13b-8k-no-rlhf-test).
|
23 |
|
24 |
+
[Kaio Ken's SuperHOT 13B LoRA](https://huggingface.co/kaiokendev/superhot-13b-8k-no-rlhf-test) is merged on to the base model, and then 8K context can be achieved during inference by using `trust_remote_code=True`.
|
25 |
|
26 |
Note that `config.json` has been set to a sequence length of 8192. This can be modified to 4096 if you want to try with a smaller sequence length.
|
27 |
|