jartine commited on
Commit
3183783
1 Parent(s): 052c7c7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +24 -0
README.md CHANGED
@@ -67,6 +67,30 @@ computation speed (flops) so simpler quants help.
67
 
68
  Note: BF16 is currently only supported on CPU.
69
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
70
  ---
71
 
72
  ## Model Details
 
67
 
68
  Note: BF16 is currently only supported on CPU.
69
 
70
+ ## Hardware Choices
71
+
72
+ Any Macbook with 32GB should be able to run
73
+ Meta-Llama-3-70B-Instruct.Q2\_K.llamafile which I uploaded a few minutes
74
+ ago. It's smart enough to solve math riddles, but at this level of
75
+ quantization you should expect hallucinations.
76
+
77
+ If you want to run Q4\_0 you'll probably be able to squeeze it on a
78
+ $3,999.00 Macbook Pro M3 Max w/ 48GB of RAM.
79
+
80
+ If you want to run Q5\_K\_M or or Q8\_0 the best choice is probably Mac
81
+ Studio. I have an Apple M2 Ultra w/ 24-core CPU, 60-core GPU, 128GB RAM.
82
+ It cost me $8000 with the monitor. If I run
83
+ Meta-Llama-3-70B-Instruct.Q4\_0.llamafile then I get 14 tok/sec (prompt
84
+ eval is 82 tok/sec) thanks to the Metal GPU.
85
+
86
+ You could alternatively go on vast.ai and rent a system with 4x RTX
87
+ 4090's for a few bucks an hour. That'll run 70b. Or you could build your
88
+ own, but the graphics cards alone will cost $10k+.
89
+
90
+ AMD Threadripper Pro 7995WX ($10k) does a good job too. I get 5.9
91
+ tok/sec eval with Q4\_0 and 49 tok/sec prompt. If I use F16 weights then
92
+ prompt eval goes 65 tok/sec.
93
+
94
  ---
95
 
96
  ## Model Details