File size: 2,618 Bytes
69f49fd c5dac99 69f49fd c94e830 05aa199 c94e830 05aa199 c5dac99 05aa199 c5dac99 e3492bc c5dac99 fd34c15 c5dac99 96aa9b7 c5dac99 346e4be c5dac99 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 |
---
license: apache-2.0
datasets:
- pankajmathur/WizardLM_Orca
language:
- en
pipeline_tag: text-generation
---
## Mistral 7b Reverse Instruct
This LoRA Adapter is fine tuned to reverse engineer the original prompt of a given LLM output/response.
- base_model: mistralai/Mistral-7B-v0.1 (=checkpoint-v1)
- base_model: mistralai/Mistral-7B-v0.2 (>=checkpoint-v2)
For convinience the latest model export is provided under [/latest_model_export](https://huggingface.co/Philipp-Sc/mistral-7b-reverse-instruct/tree/main/latest_model_export) as well as gguf quantized versions under [/latest_ggml_models](https://huggingface.co/Philipp-Sc/mistral-7b-reverse-instruct/tree/main/latest_ggml_models)
## Response Format
"[INST]\n### System:\n{system}\n### Instruction:\n{instruction}\n[/INST]\n"
- Grammar File: [inst_format.gbnf](https://huggingface.co/Philipp-Sc/mistral-7b-reverse-instruct/blob/main/inst_format.gbnf)
## Prompt Template
"\n### System:\nYou craft instructions for generating the given output through reverse engineering.\n### Instruction:\nDecipher the steps used to produce the given output and articulate a refined set of instructions (System & Instruction).\n### OUTPUT:\n {output}"
(use the template without the " ")
## Training Dataset
About 21k items of the following datasets were used. (mostly coding-like tasks were removed)
```bash
wget https://raw.githubusercontent.com/Instruction-Tuning-with-GPT-4/GPT-4-LLM/main/data/alpaca_gpt4_data.json
wget https://raw.githubusercontent.com/teknium1/GPTeacher/main/Roleplay%20Supplemental/roleplay-instruct-v2.1.json
wget https://huggingface.co/datasets/pankajmathur/WizardLM_Orca/resolve/main/wizardlm_orca.json
```
## Training Procedure
```bash
CUDA_VISIBLE_DEVICES=0 WANDB_DISABLED=True python LLaMA-Factory/src/train_bash.py \
--stage sft \
--model_name_or_path model_name_or_path \
--checkpoint_dir checkpoint_dir \
--flash_attn \
--shift_attn \
--neftune_noise_alpha 5 \
--do_train \
--dataset default \
--template vanilla \
--finetuning_type lora \
--lora_target q_proj,v_proj \
--output_dir path_to_sft_checkpoint \
--overwrite_cache \
--per_device_train_batch_size 1 \
--gradient_accumulation_steps 1 \
--lr_scheduler_type cosine \
--logging_steps 10 \
--save_steps 100 \
--learning_rate 5e-5 \
--num_train_epochs 3.0 \
--plot_loss \
--fp16 \
--overwrite_output_dir \
--cutoff_len 2048 \
--quantization_bit 4
```
## Training Time
- v1: ~12h on Kaggle's P100 GPU
- v2: >30h on Kaggle's T4 x2
### Framework versions
- LLaMA-Factory |