TheBloke commited on
Commit
b3dad87
1 Parent(s): 7efe9c6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -1
README.md CHANGED
@@ -40,6 +40,12 @@ tags:
40
 
41
  This repo contains GGML format model files for [NousResearch's Nous Hermes Llama2 70B](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-70b).
42
 
 
 
 
 
 
 
43
  GPU acceleration is now available for Llama 2 70B GGML files, with both CUDA (NVidia) and Metal (macOS). The following clients/libraries are known to work with these files, including with GPU acceleration:
44
  * [llama.cpp](https://github.com/ggerganov/llama.cpp), commit `e76d630` and later.
45
  * [text-generation-webui](https://github.com/oobabooga/text-generation-webui), the most widely used web UI.
@@ -68,7 +74,7 @@ GPU acceleration is now available for Llama 2 70B GGML files, with both CUDA (NV
68
  <!-- compatibility_ggml start -->
69
  ## Compatibility
70
 
71
- ### Requires llama.cpp [commit `e76d630`](https://github.com/ggerganov/llama.cpp/commit/e76d630df17e235e6b9ef416c45996765d2e36fb) or later.
72
 
73
  Or one of the other tools and libraries listed above.
74
 
@@ -142,6 +148,10 @@ sudo apt update -y && sudo apt install 7zip
142
 
143
  ## How to run in `llama.cpp`
144
 
 
 
 
 
145
  I use the following command line; adjust for your tastes and needs:
146
 
147
  ```
 
40
 
41
  This repo contains GGML format model files for [NousResearch's Nous Hermes Llama2 70B](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-70b).
42
 
43
+ ### Important note regarding GGML files.
44
+
45
+ The GGML format has now been superseded by GGUF. As of August 21st 2023, [llama.cpp](https://github.com/ggerganov/llama.cpp) no longer supports GGML models. Third party clients and libraries are expected to still support it for a time, but many may also drop support.
46
+
47
+ ### About GGML
48
+
49
  GPU acceleration is now available for Llama 2 70B GGML files, with both CUDA (NVidia) and Metal (macOS). The following clients/libraries are known to work with these files, including with GPU acceleration:
50
  * [llama.cpp](https://github.com/ggerganov/llama.cpp), commit `e76d630` and later.
51
  * [text-generation-webui](https://github.com/oobabooga/text-generation-webui), the most widely used web UI.
 
74
  <!-- compatibility_ggml start -->
75
  ## Compatibility
76
 
77
+ ### Compatible with [commit `e76d630`](https://github.com/ggerganov/llama.cpp/commit/e76d630df17e235e6b9ef416c45996765d2e36fb) until August 21st 2023.
78
 
79
  Or one of the other tools and libraries listed above.
80
 
 
148
 
149
  ## How to run in `llama.cpp`
150
 
151
+ Make sure you are using `llama.cpp` from commit [dadbed99e65252d79f81101a392d0d6497b86caa](https://github.com/ggerganov/llama.cpp/commit/dadbed99e65252d79f81101a392d0d6497b86caa) or earlier.
152
+
153
+ For compatibility with latest llama.cpp, please use GGUF files instead.
154
+
155
  I use the following command line; adjust for your tastes and needs:
156
 
157
  ```