Update README.md
Browse files
README.md
CHANGED
@@ -6,50 +6,15 @@ license: cc-by-nc-4.0
|
|
6 |
|
7 |
This is a fine-tuned 7b parameter LlaMa model, using completely synthetic training data created by https://github.com/jondurbin/airoboros
|
8 |
|
|
|
|
|
|
|
|
|
9 |
### Training data
|
10 |
|
11 |
-
|
12 |
-
|
13 |
-
The jailbreak prompt I used is the default prompt in the python code when using the `--uncensored` flag: https://github.com/jondurbin/airoboros/blob/main/airoboros/self_instruct.py#L39
|
14 |
-
|
15 |
-
I also did a few passes of manually cleanup to remove some bad prompts, but mostly I left the data as-is. Initially, the model was fairly bad at math/extrapolation, closed question-answering (heavy hallucination), and coding, so I did one more fine tuning pass with additional synthetic instructions aimed at those types of problems.
|
16 |
-
|
17 |
-
Both the initial instructions and final-pass fine-tuning instructions will be published soon.
|
18 |
-
|
19 |
-
### Fine-tuning method
|
20 |
-
|
21 |
-
I used the excellent [FastChat](https://github.com/lm-sys/FastChat) module, running with:
|
22 |
-
|
23 |
-
```
|
24 |
-
torchrun --nproc_per_node=8 --master_port=20001 /workspace/FastChat/fastchat/train/train_mem.py \
|
25 |
-
--model_name_or_path /workspace/llama-7b \
|
26 |
-
--data_path /workspace/as_conversations.json \
|
27 |
-
--bf16 True \
|
28 |
-
--output_dir /workspace/airoboros-7b \
|
29 |
-
--num_train_epochs 3 \
|
30 |
-
--per_device_train_batch_size 24 \
|
31 |
-
--per_device_eval_batch_size 24 \
|
32 |
-
--gradient_accumulation_steps 2 \
|
33 |
-
--evaluation_strategy "steps" \
|
34 |
-
--eval_steps 1000 \
|
35 |
-
--save_strategy "steps" \
|
36 |
-
--save_steps 1000 \
|
37 |
-
--save_total_limit 10 \
|
38 |
-
--learning_rate 2e-5 \
|
39 |
-
--weight_decay 0. \
|
40 |
-
--warmup_ratio 0.04 \
|
41 |
-
--lr_scheduler_type "cosine" \
|
42 |
-
--logging_steps 1 \
|
43 |
-
--fsdp "full_shard auto_wrap" \
|
44 |
-
--fsdp_transformer_layer_cls_to_wrap 'LlamaDecoderLayer' \
|
45 |
-
--tf32 True \
|
46 |
-
--model_max_length 2048 \
|
47 |
-
--gradient_checkpointing True \
|
48 |
-
--lazy_preprocess True
|
49 |
-
```
|
50 |
-
|
51 |
-
This ran on 8x nvidia 80gb a100's for about 17 hours.
|
52 |
|
|
|
53 |
|
54 |
### Usage and License Notices
|
55 |
|
|
|
6 |
|
7 |
This is a fine-tuned 7b parameter LlaMa model, using completely synthetic training data created by https://github.com/jondurbin/airoboros
|
8 |
|
9 |
+
__*I don't recommend using this model! The outputs aren't particularly great, and it may contain "harmful" data due to jailbreak*__
|
10 |
+
|
11 |
+
Please see one of the updated airoboros models for a much better experience.
|
12 |
+
|
13 |
### Training data
|
14 |
|
15 |
+
This was an experiment to see if a "jailbreak" prompt could be used to generate a broader range of data that would otherwise have been filtered by OpenAI's alignment efforts.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
|
17 |
+
The jailbreak did indeed work with a high success rate, and caused OpenAI to generate a broader range of topics and fewer refusals to answer questions/instructions of sensitive topics.
|
18 |
|
19 |
### Usage and License Notices
|
20 |
|