ping98k commited on
Commit
33dbe0e
1 Parent(s): c70c42d

Upload model

Browse files
Files changed (1) hide show
  1. README.md +32 -4
README.md CHANGED
@@ -2,7 +2,7 @@
2
  library_name: peft
3
  ---
4
 
5
- `facebook/xglm-7.5B` lora fine tune with `Thaweewat/alpaca-cleaned-52k-th`
6
 
7
 
8
  template
@@ -12,9 +12,24 @@ input
12
  ### Answer:
13
  ```
14
 
15
- ---
16
- library_name: peft
17
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  ## Training procedure
19
 
20
 
@@ -42,8 +57,21 @@ The following `bitsandbytes` quantization config was used during training:
42
  - bnb_4bit_use_double_quant: False
43
  - bnb_4bit_compute_dtype: float32
44
 
 
 
 
 
 
 
 
 
 
 
 
 
45
 
46
  ### Framework versions
47
 
48
  - PEFT 0.6.0.dev0
49
  - PEFT 0.6.0.dev0
 
 
2
  library_name: peft
3
  ---
4
 
5
+ lora fine tune `facebook/xglm-7.5B` with `Thaweewat/alpaca-cleaned-52k-th`
6
 
7
 
8
  template
 
12
  ### Answer:
13
  ```
14
 
15
+ ```
16
+ peft_config = LoraConfig(
17
+ r=64,
18
+ lora_alpha=128,
19
+ lora_dropout=0.05,
20
+ bias="none",
21
+ task_type="CAUSAL_LM",
22
+ target_modules=[
23
+ "q_proj",
24
+ "k_proj",
25
+ "v_proj",
26
+ "out_proj",
27
+ "fc1",
28
+ "fc2",
29
+ ]
30
+ )
31
+ ```
32
+
33
  ## Training procedure
34
 
35
 
 
57
  - bnb_4bit_use_double_quant: False
58
  - bnb_4bit_compute_dtype: float32
59
 
60
+ The following `bitsandbytes` quantization config was used during training:
61
+ - quant_method: bitsandbytes
62
+ - load_in_8bit: False
63
+ - load_in_4bit: True
64
+ - llm_int8_threshold: 6.0
65
+ - llm_int8_skip_modules: None
66
+ - llm_int8_enable_fp32_cpu_offload: False
67
+ - llm_int8_has_fp16_weight: False
68
+ - bnb_4bit_quant_type: fp4
69
+ - bnb_4bit_use_double_quant: False
70
+ - bnb_4bit_compute_dtype: float32
71
+
72
 
73
  ### Framework versions
74
 
75
  - PEFT 0.6.0.dev0
76
  - PEFT 0.6.0.dev0
77
+ - PEFT 0.6.0.dev0