habanoz commited on
Commit
586c223
1 Parent(s): 4ff98b9

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +62 -0
README.md ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - OpenAssistant/oasst_top1_2023-08-25
5
+ language:
6
+ - en
7
+ pipeline_tag: text-generation
8
+ base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-955k-token-2T
9
+ ---
10
+
11
+ TinyLlama/TinyLlama-1.1B-intermediate-step-955k-token-2T finetuned using OpenAssistant/oasst_top1_2023-08-25 dataset.
12
+
13
+ Trained for 5 epochs using Qlora. Adapter is merged.
14
+
15
+ SFT code:
16
+ https://github.com/habanoz/qlora.git
17
+
18
+ Command used:
19
+ ```bash
20
+ accelerate launch $BASE_DIR/qlora/train.py \
21
+ --model_name_or_path $BASE_MODEL \
22
+ --working_dir $BASE_DIR/$OUTPUT_NAME-checkpoints \
23
+ --output_dir $BASE_DIR/$OUTPUT_NAME-peft \
24
+ --merged_output_dir $BASE_DIR/$OUTPUT_NAME \
25
+ --final_output_dir $BASE_DIR/$OUTPUT_NAME-final \
26
+ --num_train_epochs 5 \
27
+ --logging_steps 1 \
28
+ --save_strategy steps \
29
+ --save_steps 75 \
30
+ --save_total_limit 2 \
31
+ --data_seed 11422 \
32
+ --evaluation_strategy steps \
33
+ --per_device_eval_batch_size 4 \
34
+ --eval_dataset_size 0.01 \
35
+ --eval_steps 75 \
36
+ --max_new_tokens 1024 \
37
+ --dataloader_num_workers 3 \
38
+ --logging_strategy steps \
39
+ --do_train \
40
+ --do_eval \
41
+ --lora_r 64 \
42
+ --lora_alpha 16 \
43
+ --lora_modules all \
44
+ --bits 4 \
45
+ --double_quant \
46
+ --quant_type nf4 \
47
+ --lr_scheduler_type constant \
48
+ --dataset oasst1-top1 \
49
+ --dataset_format oasst1 \
50
+ --model_max_len 1024 \
51
+ --per_device_train_batch_size 4 \
52
+ --gradient_accumulation_steps 4 \
53
+ --learning_rate 1e-5 \
54
+ --adam_beta2 0.999 \
55
+ --max_grad_norm 0.3 \
56
+ --lora_dropout 0.0 \
57
+ --weight_decay 0.0 \
58
+ --seed 11422 \
59
+ --gradient_checkpointing \
60
+ --use_flash_attention_2 \
61
+ --ddp_find_unused_parameters False
62
+ ```