|
--- |
|
license: cc-by-nc-4.0 |
|
language: |
|
- ja |
|
pipeline_tag: text-generation |
|
--- |
|
|
|
BaseModel: RWKV-4-Pile-14B-Instruct-test5-20230329-ctx4096.pth |
|
URL: https://huggingface.co/BlinkDL/rwkv-4-pile-14b |
|
|
|
Training Code: |
|
https://github.com/Blealtan/RWKV-LM-LoRA |
|
|
|
Dataset: |
|
https://github.com/shi3z/alpaca_ja |
|
|
|
Loss: |
|
Epoch 320: loss 0.76 |
|
Epoch 500: loss 0.62 |
|
|
|
Training Prompt: |
|
python train.py \ |
|
--load_model RWKV-4-Pile-14B-Instruct-test5-20230329-ctx4096.pth \ |
|
--proj_dir out \ |
|
--data_file "train.npy" \ |
|
--data_type "numpy" \ |
|
--vocab_size 50277 \ |
|
--ctx_len 1024 \ |
|
--epoch_save 5 \ |
|
--epoch_count 100 \ |
|
--n_layer 40 \ |
|
--n_embd 5120 \ |
|
--epoch_steps 1000 --epoch_begin 0 --micro_bsz 1 --pre_ffn 0 --head_qk 0 --lr_init 1e-5 --lr_final 1e-5 --warmup_steps 0 --beta1 0.9 --beta2 0.999 --adam_eps 1e-8 --accelerator gpu --devices 1 --precision bf16 --strategy deepspeed_stage_2 --grad_cp 0 \ |
|
--lora --lora_r 8 --lora_alpha 32 --lora_dropout 0.01 |
|
|
|
Training Environment: |
|
A6000x1 + XEON 256GB RAM |