OptimizeLLM
commited on
Commit
•
7e61312
1
Parent(s):
bde709a
Update README.md
Browse files
README.md
CHANGED
@@ -35,7 +35,8 @@ The following example starts at the root of D drive and quantizes mistral's Mixt
|
|
35 |
|
36 |
## Download llama.cpp
|
37 |
Assuming you want CUDA for your NVIDIA RTX GPU(s) use the links below, or grab latest compiled executables from https://github.com/ggerganov/llama.cpp/releases
|
38 |
-
|
|
|
39 |
* https://github.com/ggerganov/llama.cpp/releases/download/b2253/cudart-llama-bin-win-cu12.2.0-x64.zip
|
40 |
* https://github.com/ggerganov/llama.cpp/releases/download/b2253/llama-b2253-bin-win-cublas-cu12.2.0-x64.zip
|
41 |
|
|
|
35 |
|
36 |
## Download llama.cpp
|
37 |
Assuming you want CUDA for your NVIDIA RTX GPU(s) use the links below, or grab latest compiled executables from https://github.com/ggerganov/llama.cpp/releases
|
38 |
+
|
39 |
+
### Latest version as of Feb 24, 2024:
|
40 |
* https://github.com/ggerganov/llama.cpp/releases/download/b2253/cudart-llama-bin-win-cu12.2.0-x64.zip
|
41 |
* https://github.com/ggerganov/llama.cpp/releases/download/b2253/llama-b2253-bin-win-cublas-cu12.2.0-x64.zip
|
42 |
|