mpasila commited on
Commit
b7bf76b
1 Parent(s): d28c8a1

Clarified quantization type

Browse files

Also added original model card

Files changed (1) hide show
  1. README.md +22 -2
README.md CHANGED
@@ -3,6 +3,26 @@ license: llama2
3
  language:
4
  - en
5
  ---
6
- This is an EXL2 quantized model in 3bpw of [jebcarter/Psyfighter-13B](https://huggingface.co/jebcarter/Psyfighter-13B) using [PIPPA](https://huggingface.co/datasets/jasonkstevens/pippa-llama2-chat/blob/refs%2Fconvert%2Fparquet/default/train/0000.parquet) as the calibration dataset. (I might reupload this or add a new branch with a different version that uses the default calibration dataset instead)
7
 
8
- Alpaca and ChatML prompt templates seem to work fine with this model.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  language:
4
  - en
5
  ---
6
+ This is an ExLlamaV2 quantized model in 3bpw of [jebcarter/Psyfighter-13B](https://huggingface.co/jebcarter/Psyfighter-13B) using [PIPPA](https://huggingface.co/datasets/jasonkstevens/pippa-llama2-chat/blob/refs%2Fconvert%2Fparquet/default/train/0000.parquet) as the calibration dataset. (I might reupload this or add a new branch with a different version that uses the default calibration dataset instead)
7
 
8
+ Alpaca and ChatML prompt templates seem to work fine with this model.
9
+
10
+ # Original Model card
11
+
12
+ ```
13
+ merge_method: task_arithmaetic
14
+ base_model: TheBloke/Llama-2-13B-fp16
15
+ models:
16
+ - model: TheBloke/Llama-2-13B-fp16
17
+ - model: KoboldAI/LLaMA2-13B-Tiefighter
18
+ parameters:
19
+ weight: 1.0
20
+ - model: chaoyi-wu/MedLLaMA_13B
21
+ parameters:
22
+ weight: 0.01
23
+ - model: Doctor-Shotgun/llama-2-13b-chat-limarp-v2-merged
24
+ parameters:
25
+ weight: 0.02
26
+ dtype: float16
27
+ ```
28
+ This model was made possible thanks to the Compute provided by the KoboldAI community.