Commit
·
4035a0c
1
Parent(s):
f78c64d
Update README.md
Browse files
README.md
CHANGED
@@ -8,7 +8,8 @@ datasets:
|
|
8 |
Fine-tuned [OpenLLaMA-7B](https://huggingface.co/openlm-research/open_llama_7b) with an uncensored/unfiltered Wizard-Vicuna conversation dataset [digitalpipelines/wizard_vicuna_70k_uncensored](https://huggingface.co/datasets/digitalpipelines/wizard_vicuna_70k_uncensored).
|
9 |
Used QLoRA for fine-tuning using the process outlined in https://georgesung.github.io/ai/qlora-ift/
|
10 |
|
11 |
-
|
|
|
12 |
|
13 |
# Prompt style
|
14 |
The model was trained with the following prompt style:
|
|
|
8 |
Fine-tuned [OpenLLaMA-7B](https://huggingface.co/openlm-research/open_llama_7b) with an uncensored/unfiltered Wizard-Vicuna conversation dataset [digitalpipelines/wizard_vicuna_70k_uncensored](https://huggingface.co/datasets/digitalpipelines/wizard_vicuna_70k_uncensored).
|
9 |
Used QLoRA for fine-tuning using the process outlined in https://georgesung.github.io/ai/qlora-ift/
|
10 |
|
11 |
+
- GPTQ quantized model can be found at [digitalpipelines/llama2_7b_chat_uncensored-GPTQ](https://huggingface.co/digitalpipelines/llama2_7b_chat_uncensored-GPTQ)
|
12 |
+
- GGML 2, 3, 4, 5, 6 and 8-bit quanitized models for CPU+GPU inference of [digitalpipelines/llama2_7b_chat_uncensored-GGML](https://huggingface.co/digitalpipelines/llama2_7b_chat_uncensored-GGML)
|
13 |
|
14 |
# Prompt style
|
15 |
The model was trained with the following prompt style:
|