BlinkDL commited on
Commit
605a997
1 Parent(s): 99835d2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -26,11 +26,11 @@ ctx_len = 1024
26
  n_layer = 32
27
  n_embd = 4096
28
 
 
 
29
  Final checkpoint: RWKV-4-Pile-7B-20221115-8047.pth : Trained on the Pile for 332B tokens.
30
  * Pile loss 1.8415
31
  * LAMBADA ppl 4.38, acc 67.18%
32
  * PIQA acc 76.06%
33
  * SC2016 acc 73.44%
34
  * Hellaswag acc_norm 65.51%
35
-
36
- I am fine-tuning it to ctx_len 4096.
 
26
  n_layer = 32
27
  n_embd = 4096
28
 
29
+ (there are ctx_len 2048 and 4096 models though they might be slightly weaker at generating short contents)
30
+
31
  Final checkpoint: RWKV-4-Pile-7B-20221115-8047.pth : Trained on the Pile for 332B tokens.
32
  * Pile loss 1.8415
33
  * LAMBADA ppl 4.38, acc 67.18%
34
  * PIQA acc 76.06%
35
  * SC2016 acc 73.44%
36
  * Hellaswag acc_norm 65.51%