File size: 995 Bytes
86f35d8
 
60e947a
 
917d507
86f35d8
60e947a
 
0ca321a
60e947a
 
 
 
72d9c90
 
 
bf4ad79
 
 
 
60e947a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
917d507
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
---
license: cc-by-nc-4.0
language:
- ja
pipeline_tag: text-generation
---

BaseModel: RWKV-4-Pile-14B-Instruct-test5-20230329-ctx4096.pth
URL: https://huggingface.co/BlinkDL/rwkv-4-pile-14b

Training Code:
https://github.com/Blealtan/RWKV-LM-LoRA

Dataset:
https://github.com/shi3z/alpaca_ja

Loss:
Epoch 320: loss 0.76
Epoch 500: loss 0.62

Training Prompt:
python train.py \
  --load_model RWKV-4-Pile-14B-Instruct-test5-20230329-ctx4096.pth \
  --proj_dir out \
  --data_file "train.npy" \
  --data_type "numpy" \
  --vocab_size 50277 \
  --ctx_len 1024 \
  --epoch_save 5 \
  --epoch_count 100 \
  --n_layer 40 \
  --n_embd 5120 \
  --epoch_steps 1000 --epoch_begin 0  --micro_bsz 1 --pre_ffn 0 --head_qk 0 --lr_init 1e-5 --lr_final 1e-5 --warmup_steps 0 --beta1 0.9 --beta2 0.999 --adam_eps 1e-8 --accelerator gpu --devices 1 --precision bf16 --strategy deepspeed_stage_2 --grad_cp 0 \
  --lora --lora_r 8 --lora_alpha 32 --lora_dropout 0.01

Training Environment:
A6000x1 + XEON 256GB RAM