TheBloke commited on
Commit
9bb0193
1 Parent(s): 79dc76e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -21,7 +21,7 @@ license: other
21
 
22
  These files are GPTQ 4bit model files for [Eric Hartford's WizardLM 13B V1.0 Uncensored](https://huggingface.co/ehartford/WizardLM-13B-V1.0-Uncensored) merged with [Kaio Ken's SuperHOT 8K](https://huggingface.co/kaiokendev/superhot-13b-8k-no-rlhf-test).
23
 
24
- [Kaio Ken's SuperHOT 30B LoRA](https://huggingface.co/kaiokendev/superhot-30b-8k-no-rlhf-test) is merged on to the base model, and then 8K context can be achieved during inference by using `trust_remote_code=True`.
25
 
26
  Note that `config.json` has been set to a sequence length of 8192. This can be modified to 4096 if you want to try with a smaller sequence length.
27
 
 
21
 
22
  These files are GPTQ 4bit model files for [Eric Hartford's WizardLM 13B V1.0 Uncensored](https://huggingface.co/ehartford/WizardLM-13B-V1.0-Uncensored) merged with [Kaio Ken's SuperHOT 8K](https://huggingface.co/kaiokendev/superhot-13b-8k-no-rlhf-test).
23
 
24
+ [Kaio Ken's SuperHOT 13B LoRA](https://huggingface.co/kaiokendev/superhot-13b-8k-no-rlhf-test) is merged on to the base model, and then 8K context can be achieved during inference by using `trust_remote_code=True`.
25
 
26
  Note that `config.json` has been set to a sequence length of 8192. This can be modified to 4096 if you want to try with a smaller sequence length.
27