8bit-coder commited on
Commit
9f6c26e
1 Parent(s): 46bf0a0

updated training instructions

Browse files
training_files/full-training-instructions.txt CHANGED
@@ -33,17 +33,19 @@ cd ..
33
 
34
  torchrun --nproc_per_node=8 --master_port=3045 ./stanford_alpaca/train.py --model_name_or_path ./llama-7b-hf --data_path ./alpaca-7b-nativeEnhanced/training_files/alpaca-megaset-fixed.json --fp16 True --output_dir ./output_7b --num_train_epochs 3 --per_device_train_batch_size 2 --per_device_eval_batch_size 2 --gradient_accumulation_steps 16 --evaluation_strategy "no" --save_strategy "steps" --save_steps 200 --learning_rate 2e-5 --weight_decay 0. --warmup_ratio 0.03 --lr_scheduler_type "cosine" --logging_steps 1 --fsdp "full_shard auto_wrap" --fsdp_transformer_layer_cls_to_wrap 'LlamaDecoderLayer' --tf32 True
35
 
36
- # now, make sure with nano that script1.py has proper paths to everything
37
 
38
  pip install -q datasets loralib sentencepiece
39
  pip install bitsandbytes
40
 
41
- python script1.py
42
 
43
  git clone https://github.com/antimatter15/alpaca.cpp
44
 
45
  cd alpaca.cpp
 
46
  mkdir models
 
47
  cd ..
48
 
49
  mv consolidated.01.pth ./alpaca.cpp/models/consolidated.00.pth
 
33
 
34
  torchrun --nproc_per_node=8 --master_port=3045 ./stanford_alpaca/train.py --model_name_or_path ./llama-7b-hf --data_path ./alpaca-7b-nativeEnhanced/training_files/alpaca-megaset-fixed.json --fp16 True --output_dir ./output_7b --num_train_epochs 3 --per_device_train_batch_size 2 --per_device_eval_batch_size 2 --gradient_accumulation_steps 16 --evaluation_strategy "no" --save_strategy "steps" --save_steps 200 --learning_rate 2e-5 --weight_decay 0. --warmup_ratio 0.03 --lr_scheduler_type "cosine" --logging_steps 1 --fsdp "full_shard auto_wrap" --fsdp_transformer_layer_cls_to_wrap 'LlamaDecoderLayer' --tf32 True
35
 
36
+ # now, make sure with nano that convert-hf-to-pth-16b.py has proper paths to everything
37
 
38
  pip install -q datasets loralib sentencepiece
39
  pip install bitsandbytes
40
 
41
+ python convert-hf-to-pth-16b.py
42
 
43
  git clone https://github.com/antimatter15/alpaca.cpp
44
 
45
  cd alpaca.cpp
46
+
47
  mkdir models
48
+
49
  cd ..
50
 
51
  mv consolidated.01.pth ./alpaca.cpp/models/consolidated.00.pth