NouRed commited on
Commit
a5f6fa1
1 Parent(s): 98be240

Upload model

Browse files
Files changed (2) hide show
  1. README.md +1 -10
  2. adapter_config.json +5 -5
README.md CHANGED
@@ -1,14 +1,6 @@
1
  ---
2
  library_name: peft
3
- license: mit
4
- datasets:
5
- - yahma/alpaca-cleaned
6
- pipeline_tag: text-generation
7
  ---
8
-
9
- ## Quantized LLaMA2 Alpaca
10
- **quantized-llama2-alpaca** is a fine-tuned version of the [LLaMA2](https://ai.meta.com/research/publications/llama-2-open-foundation-and-fine-tuned-chat-models/) ([Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b)) model on the [Alpaca](https://github.com/tatsu-lab/stanford_alpaca) dataset using [QLoRA](https://arxiv.org/abs/2305.14314).
11
-
12
  ## Training procedure
13
 
14
 
@@ -22,8 +14,7 @@ The following `bitsandbytes` quantization config was used during training:
22
  - bnb_4bit_quant_type: nf4
23
  - bnb_4bit_use_double_quant: True
24
  - bnb_4bit_compute_dtype: bfloat16
25
-
26
  ### Framework versions
27
 
28
 
29
- - PEFT 0.4.0
 
1
  ---
2
  library_name: peft
 
 
 
 
3
  ---
 
 
 
 
4
  ## Training procedure
5
 
6
 
 
14
  - bnb_4bit_quant_type: nf4
15
  - bnb_4bit_use_double_quant: True
16
  - bnb_4bit_compute_dtype: bfloat16
 
17
  ### Framework versions
18
 
19
 
20
+ - PEFT 0.4.0
adapter_config.json CHANGED
@@ -14,13 +14,13 @@
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
- "v_proj",
 
18
  "gate_proj",
19
- "down_proj",
20
- "o_proj",
21
  "up_proj",
22
- "k_proj",
23
- "q_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
 
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
+ "q_proj",
18
+ "k_proj",
19
  "gate_proj",
20
+ "v_proj",
 
21
  "up_proj",
22
+ "o_proj",
23
+ "down_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }