OptimizeLLM commited on
Commit
7e61312
1 Parent(s): bde709a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -35,7 +35,8 @@ The following example starts at the root of D drive and quantizes mistral's Mixt
35
 
36
  ## Download llama.cpp
37
  Assuming you want CUDA for your NVIDIA RTX GPU(s) use the links below, or grab latest compiled executables from https://github.com/ggerganov/llama.cpp/releases
38
- Latest version as of Feb 24, 2024:
 
39
  * https://github.com/ggerganov/llama.cpp/releases/download/b2253/cudart-llama-bin-win-cu12.2.0-x64.zip
40
  * https://github.com/ggerganov/llama.cpp/releases/download/b2253/llama-b2253-bin-win-cublas-cu12.2.0-x64.zip
41
 
 
35
 
36
  ## Download llama.cpp
37
  Assuming you want CUDA for your NVIDIA RTX GPU(s) use the links below, or grab latest compiled executables from https://github.com/ggerganov/llama.cpp/releases
38
+
39
+ ### Latest version as of Feb 24, 2024:
40
  * https://github.com/ggerganov/llama.cpp/releases/download/b2253/cudart-llama-bin-win-cu12.2.0-x64.zip
41
  * https://github.com/ggerganov/llama.cpp/releases/download/b2253/llama-b2253-bin-win-cublas-cu12.2.0-x64.zip
42