--- library_name: peft --- Authors: Forrest Kim, Joe Damisch, He Shi ## Training procedure A BERT classifier was trained on the Stereoset dataset (https://huggingface.co/datasets/stereoset). Only the racial and professional prompts were used for training the classifier due to covariate imbalance. LLaMA 2 7B chat (https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) was used to prompt a random sample of the Stereoset dataset prompts, as well as, from a manually curated list of leading prompts (What is wrong with doctors?, Why are programmers so skinny?, etc.) The outcomes of these prompts were then classified using the BERT Classifier and the resulting 'unbiased' context, prompt pairings were used to fine-tune the LLaMA 2 chat model using the PEFT LORA library. PEFT: The following bitsandbytes quantization config was used during training: load_in_8bit: False load_in_4bit: True llm_int8_threshold: 6.0 llm_int8_skip_modules: None llm_int8_enable_fp32_cpu_offload: False llm_int8_has_fp16_weight: False bnb_4bit_quant_type: fp4 bnb_4bit_use_double_quant: False bnb_4bit_compute_dtype: float32 This ran on for 1000 steps -- 1 epochs on a stereotype dataset. Training took ~1 hour on a 4090. ### Framework versions - PEFT 0.5.0.dev0