jondurbin commited on
Commit
2579ee9
1 Parent(s): 7ea67f8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -41
README.md CHANGED
@@ -6,50 +6,15 @@ license: cc-by-nc-4.0
6
 
7
  This is a fine-tuned 7b parameter LlaMa model, using completely synthetic training data created by https://github.com/jondurbin/airoboros
8
 
 
 
 
 
9
  ### Training data
10
 
11
- I used a jailbreak prompt to generate the synthetic instructions, which resulted in some training data that would likely be censored by other models, such as how-to prompts about synthesizing drugs, making homemade flamethrowers, etc. Mind you, this is all generated by ChatGPT, not me. My goal was to simply test some of the capabilities of ChatGPT when unfiltered (as much as possible), and not to intentionally produce any harmful/dangerous/etc. content.
12
-
13
- The jailbreak prompt I used is the default prompt in the python code when using the `--uncensored` flag: https://github.com/jondurbin/airoboros/blob/main/airoboros/self_instruct.py#L39
14
-
15
- I also did a few passes of manually cleanup to remove some bad prompts, but mostly I left the data as-is. Initially, the model was fairly bad at math/extrapolation, closed question-answering (heavy hallucination), and coding, so I did one more fine tuning pass with additional synthetic instructions aimed at those types of problems.
16
-
17
- Both the initial instructions and final-pass fine-tuning instructions will be published soon.
18
-
19
- ### Fine-tuning method
20
-
21
- I used the excellent [FastChat](https://github.com/lm-sys/FastChat) module, running with:
22
-
23
- ```
24
- torchrun --nproc_per_node=8 --master_port=20001 /workspace/FastChat/fastchat/train/train_mem.py \
25
- --model_name_or_path /workspace/llama-7b \
26
- --data_path /workspace/as_conversations.json \
27
- --bf16 True \
28
- --output_dir /workspace/airoboros-7b \
29
- --num_train_epochs 3 \
30
- --per_device_train_batch_size 24 \
31
- --per_device_eval_batch_size 24 \
32
- --gradient_accumulation_steps 2 \
33
- --evaluation_strategy "steps" \
34
- --eval_steps 1000 \
35
- --save_strategy "steps" \
36
- --save_steps 1000 \
37
- --save_total_limit 10 \
38
- --learning_rate 2e-5 \
39
- --weight_decay 0. \
40
- --warmup_ratio 0.04 \
41
- --lr_scheduler_type "cosine" \
42
- --logging_steps 1 \
43
- --fsdp "full_shard auto_wrap" \
44
- --fsdp_transformer_layer_cls_to_wrap 'LlamaDecoderLayer' \
45
- --tf32 True \
46
- --model_max_length 2048 \
47
- --gradient_checkpointing True \
48
- --lazy_preprocess True
49
- ```
50
-
51
- This ran on 8x nvidia 80gb a100's for about 17 hours.
52
 
 
53
 
54
  ### Usage and License Notices
55
 
 
6
 
7
  This is a fine-tuned 7b parameter LlaMa model, using completely synthetic training data created by https://github.com/jondurbin/airoboros
8
 
9
+ __*I don't recommend using this model! The outputs aren't particularly great, and it may contain "harmful" data due to jailbreak*__
10
+
11
+ Please see one of the updated airoboros models for a much better experience.
12
+
13
  ### Training data
14
 
15
+ This was an experiment to see if a "jailbreak" prompt could be used to generate a broader range of data that would otherwise have been filtered by OpenAI's alignment efforts.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
 
17
+ The jailbreak did indeed work with a high success rate, and caused OpenAI to generate a broader range of topics and fewer refusals to answer questions/instructions of sensitive topics.
18
 
19
  ### Usage and License Notices
20