reissbaker commited on
Commit
97e01d4
·
1 Parent(s): a2f8007

Add README

Browse files
Files changed (1) hide show
  1. README.md +43 -0
README.md ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: meta-llama/Meta-Llama-3.1-70B-Instruct
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ This LoRA adapter was extracted from
9
+ [mlabonne/Meta-Llama-3.1-70B-Instruct-lorablated](https://huggingface.co/mlabonne/Meta-Llama-3.1-70B-Instruct-lorablated)
10
+ and uses
11
+ [meta-llama/Llama-3.1-70B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-70B-Instruct)
12
+ as a base.
13
+
14
+ ## Model Details
15
+
16
+ The model was extracted by running
17
+ [mlabonne/harmful_behaviors](https://huggingface.co/datasets/mlabonne/harmful_behaviors)
18
+ and the user prompts (but not assistant responses or system messages) from
19
+ [Guilherme34/uncensor](https://huggingface.co/datasets/Guilherme34/uncensor)
20
+ through the original abliterated model to generate a dataset of
21
+ prompt/completion pairs, and was trained for 2 epochs on a 8xA100s with Axolotl
22
+ using FSDP. Since the original abliterated model isn't perfect at avoiding
23
+ refusals, the dataset was cleaned to remove the few refusals generated prior to
24
+ training.
25
+
26
+ ### Model Description
27
+
28
+ - **Developed by:** @reissbaker
29
+ - **Funded by:** Synthetic Lab
30
+ - **License:** Apache 2.0
31
+ - **Finetuned from model:** Llama 3.1 70B Instruct
32
+
33
+ ## How to Get Started with the Model
34
+
35
+ Run the model with one click on [glhf.chat](https://glhf.chat).
36
+
37
+
38
+ #### Training Hyperparameters
39
+
40
+ * BF16 mixed-precision
41
+ * 4e-4 LR
42
+ * Linear LR schedule
43
+ * Fused AdamW optimizer