Jared Van Bortel commited on
Commit
a4ff1a1
1 Parent(s): d7e1f72

README: add note about llama.cpp version

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -25,6 +25,8 @@ This repo contains llama.cpp-compatible files for [nomic-embed-text-v1](https://
25
 
26
  llama.cpp will default to 2048 tokens of context with these files. To use the full 8192 tokens that Nomic Embed is benchmarked on, you will have to choose a context extension method. The original model uses Dynamic NTK-Aware RoPE scaling, but that is not currently available in llama.cpp. A combination of YaRN and linear scaling is an acceptable substitute.
27
 
 
 
28
  ## Example `llama.cpp` Command
29
 
30
  Compute a single embedding:
 
25
 
26
  llama.cpp will default to 2048 tokens of context with these files. To use the full 8192 tokens that Nomic Embed is benchmarked on, you will have to choose a context extension method. The original model uses Dynamic NTK-Aware RoPE scaling, but that is not currently available in llama.cpp. A combination of YaRN and linear scaling is an acceptable substitute.
27
 
28
+ These files were converted and quantized with llama.cpp commit [6c00a0669](https://github.com/ggerganov/llama.cpp/commit/6c00a066928b0475b865a2e3e709e2166e02d548).
29
+
30
  ## Example `llama.cpp` Command
31
 
32
  Compute a single embedding: