forki commited on
Commit
8a61628
1 Parent(s): 875069b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -1
README.md CHANGED
@@ -3,7 +3,29 @@ library_name: peft
3
  ---
4
  ## Training procedure
5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  ### Framework versions
7
 
8
 
9
- - PEFT 0.5.0.dev0
 
3
  ---
4
  ## Training procedure
5
 
6
+ A BERT classifier was trained on the Stereoset dataset (https://huggingface.co/datasets/stereoset). Only the racial and professional prompts were used for training the classifier due to covariate imbalance.
7
+
8
+ LLaMA 2 7B chat (https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) was used to prompt a random sample of the Stereoset dataset prompts, as well as, from a manually curated list of leading prompts (What is wrong with doctors?, Why are programmers so skinny?, etc.)
9
+
10
+ The outcomes of these prompts were then classified using the BERT Classifier and the resulting 'unbiased' context, prompt pairings were used to fine-tune the LLaMA 2 chat model using the PEFT LORA library.
11
+
12
+ PEFT:
13
+
14
+ The following bitsandbytes quantization config was used during training:
15
+
16
+ load_in_8bit: False
17
+ load_in_4bit: True
18
+ llm_int8_threshold: 6.0
19
+ llm_int8_skip_modules: None
20
+ llm_int8_enable_fp32_cpu_offload: False
21
+ llm_int8_has_fp16_weight: False
22
+ bnb_4bit_quant_type: fp4
23
+ bnb_4bit_use_double_quant: False
24
+ bnb_4bit_compute_dtype: float32
25
+
26
+ This ran on for 1000 steps -- 1 epochs on a stereotype dataset. Training took ~1 hour on a 4090.
27
+
28
  ### Framework versions
29
 
30
 
31
+ - PEFT 0.5.0.dev0