tricktreat commited on
Commit
f2bdf5c
1 Parent(s): a4a02e4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -4
README.md CHANGED
@@ -7,7 +7,7 @@ license: apache-2.0
7
  # Training
8
 
9
  ```
10
- deepspeed --include=node-0:3 --master_port=12001 sft_prompt_tuning.py --deepspeed dp_zero0.json \
11
  --model_name_or_path="guanaco_Llama-2-7b-chat-hf_freeze_embed_tokens_q_v_proj" \
12
  --dataset_name="timdettmers/openassistant-guanaco" \
13
  --dataset_text_field="text" \
@@ -15,13 +15,16 @@ deepspeed --include=node-0:3 --master_port=12001 sft_prompt_tuning.py --deepspee
15
  --learning_rate=1e-5 \
16
  --per_device_train_batch_size=32 \
17
  --gradient_accumulation_steps=4 \
18
- --output_dir="guanaco_Llama-2-7b-chat-hf_freeze_embed_tokens_q_v_projs_prompttuning" \
19
  --logging_steps=1 \
20
  --num_train_epochs=15 \
21
  --max_steps=-1 \
22
- --save_steps=0.3 \
23
  --gradient_checkpointing \
24
- --fp16
 
 
 
 
25
  ```
26
 
27
  # Model Card for Model ID
 
7
  # Training
8
 
9
  ```
10
+ deepspeed --include=node-0:2 sft_fix_target_modules.py --deepspeed dp_zero0.json \
11
  --model_name_or_path="guanaco_Llama-2-7b-chat-hf_freeze_embed_tokens_q_v_proj" \
12
  --dataset_name="timdettmers/openassistant-guanaco" \
13
  --dataset_text_field="text" \
 
15
  --learning_rate=1e-5 \
16
  --per_device_train_batch_size=32 \
17
  --gradient_accumulation_steps=4 \
18
+ --output_dir="guanaco_Llama-2-7b-chat-hf_freeze_embed_tokens_q_v_proj_lora" \
19
  --logging_steps=1 \
20
  --num_train_epochs=15 \
21
  --max_steps=-1 \
 
22
  --gradient_checkpointing \
23
+ --fp16 \
24
+ --save_steps=0.3 \
25
+ --use_peft \
26
+ --lora_r=64 \
27
+ --lora_alpha=16
28
  ```
29
 
30
  # Model Card for Model ID