Update README.md
Browse files
README.md
CHANGED
@@ -49,7 +49,7 @@ It uses Cosmopolitan Libc to turn LLM weights into runnable llama.cpp
|
|
49 |
binaries that run on the stock installs of six OSes for both ARM64 and
|
50 |
AMD64.
|
51 |
|
52 |
-
## About Quantization Formats
|
53 |
|
54 |
Your choice of quantization format depends on three things:
|
55 |
|
@@ -67,7 +67,7 @@ computation speed (flops) so simpler quants help.
|
|
67 |
|
68 |
Note: BF16 is currently only supported on CPU.
|
69 |
|
70 |
-
## Hardware Choices
|
71 |
|
72 |
Any Macbook with 32GB should be able to run
|
73 |
Meta-Llama-3-70B-Instruct.Q2\_K.llamafile reasonably well. At this
|
|
|
49 |
binaries that run on the stock installs of six OSes for both ARM64 and
|
50 |
AMD64.
|
51 |
|
52 |
+
## About Quantization Formats (General Advice)
|
53 |
|
54 |
Your choice of quantization format depends on three things:
|
55 |
|
|
|
67 |
|
68 |
Note: BF16 is currently only supported on CPU.
|
69 |
|
70 |
+
## Hardware Choices (LLaMA3 70B Specific)
|
71 |
|
72 |
Any Macbook with 32GB should be able to run
|
73 |
Meta-Llama-3-70B-Instruct.Q2\_K.llamafile reasonably well. At this
|