habanoz commited on
Commit
c0e3f47
1 Parent(s): 60af1b3

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +51 -0
README.md ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - OpenAssistant/oasst_top1_2023-08-25
5
+ language:
6
+ - en
7
+ pipeline_tag: text-generation
8
+ ---
9
+
10
+ TinyLlama-1.1B-intermediate-step-715k-1.5T finetuned using OpenAssistant/oasst_top1_2023-08-25 dataset.
11
+
12
+ SFT code:
13
+ https://github.com/jzhang38/TinyLlama/tree/main/sft
14
+
15
+ Command used:
16
+ ```bash
17
+ accelerate launch finetune.py \
18
+ --model_name_or_path TinyLlama/TinyLlama-1.1B-intermediate-step-715k-1.5T \
19
+ --output_dir ./output/1_5T_FT_lr1e-5_ep5_top1_2023-08-25 \
20
+ --logging_steps 10 \
21
+ --save_strategy epoch \
22
+ --data_seed 42 \
23
+ --save_total_limit 2 \
24
+ --evaluation_strategy epoch \
25
+ --eval_dataset_size 512 \
26
+ --max_eval_samples 1000 \
27
+ --per_device_eval_batch_size 1 \
28
+ --max_new_tokens 32 \
29
+ --dataloader_num_workers 3 \
30
+ --group_by_length=False \
31
+ --logging_strategy steps \
32
+ --remove_unused_columns False \
33
+ --do_train \
34
+ --do_eval \
35
+ --warmup_ratio 0.05 \
36
+ --lr_scheduler_type constant \
37
+ --dataset OpenAssistant/oasst_top1_2023-08-25 \
38
+ --dataset_format oasst1 \
39
+ --source_max_len 1 \
40
+ --target_max_len 1023 \
41
+ --per_device_train_batch_size 2 \
42
+ --gradient_accumulation_steps 8 \
43
+ --max_steps 0 \
44
+ --num_train_epochs 5 \
45
+ --learning_rate 1e-5 \
46
+ --adam_beta2 0.999 \
47
+ --max_grad_norm 1.0 \
48
+ --weight_decay 0.0 \
49
+ --seed 0 \
50
+ --trust_remote_code
51
+ ```