gururise commited on
Commit
750b68d
1 Parent(s): efafe5a

update config for 14b pre-quantized model

Browse files
Files changed (1) hide show
  1. config.py +17 -4
config.py CHANGED
@@ -5,8 +5,8 @@ quantized = {
5
  "mode": TORCH_QUANT,
6
  "runtimedtype": torch.bfloat16,
7
  "useGPU": torch.cuda.is_available(),
8
- "chunksize": 32, # larger = more accurate, but more memory
9
- "target": 100 # your gpu max size, excess vram offloaded to cpu
10
  }
11
 
12
  # UNCOMMENT TO SELECT OPTIONS
@@ -57,7 +57,20 @@ title = "RWKV-4 (1.5b Instruct Test 2)"
57
  # [File size: 28.0GB]
58
 
59
  # config = {
60
- # "path": "https://huggingface.co/BlinkDL/rwkv-4-pile-14b/resolve/main/RWKV-4-Pile-14B-20230204-7324.pth"
 
 
 
 
 
 
 
 
 
 
 
 
 
61
  # }
62
 
63
- # title = "RWKV-4 (14b)"
 
5
  "mode": TORCH_QUANT,
6
  "runtimedtype": torch.bfloat16,
7
  "useGPU": torch.cuda.is_available(),
8
+ "chunksize": 32, # larger = more accurate, but more memory (and slower)
9
+ "target": 24 # your gpu max size, excess vram offloaded to cpu
10
  }
11
 
12
  # UNCOMMENT TO SELECT OPTIONS
 
57
  # [File size: 28.0GB]
58
 
59
  # config = {
60
+ # "path": "https://huggingface.co/BlinkDL/rwkv-4-pile-14b/resolve/main/RWKV-4-Pile-14B-20230204-7324.pth",
61
+ # **quantized
62
+ # }
63
+
64
+ # title = "RWKV-4 (14b Quantized)"
65
+
66
+
67
+ # RWKV 14B pre-quantized (latest as of feb 9)
68
+ # Approximate
69
+ # [Vram usage: 15.0GB]
70
+ # [File size: 14.4GB]
71
+
72
+ # config = {
73
+ # "path": "https://huggingface.co/Hazzzardous/RWKV-8Bit/resolve/main/RWKV-4-Pile-14B-20230204-7324.pqth"
74
  # }
75
 
76
+ # title = "RWKV-4 (14b Quantized)"