Update README.md
Browse files
README.md
CHANGED
@@ -8,11 +8,24 @@ license: mit
|
|
8 |
|
9 |
Source model: https://huggingface.co/BAAI/bge-base-en-v1.5
|
10 |
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
|
17 |
# Inference
|
18 |
|
|
|
8 |
|
9 |
Source model: https://huggingface.co/BAAI/bge-base-en-v1.5
|
10 |
|
11 |
+
Quantized and unquantized embedding models for use with `llama.cpp`.
|
12 |
+
|
13 |
+
<br/>
|
14 |
+
|
15 |
+
# Files Available
|
16 |
+
|
17 |
+
<div style="width: 500px; margin: 0;">
|
18 |
+
|
19 |
+
| Filename | Quantization | Size |
|
20 |
+
|:-------- | ------------ | ---- |
|
21 |
+
| [bge-base-en-v1.5-f32.gguf](https://huggingface.co/CompendiumLabs/bge-base-en-v1.5-gguf/blob/main/bge-base-en-v1.5-f32.gguf) | F32 | 436 MB |
|
22 |
+
| [bge-base-en-v1.5-f16.gguf](https://huggingface.co/CompendiumLabs/bge-base-en-v1.5-gguf/blob/main/bge-base-en-v1.5-f16.gguf) | F16 | 219 MB |
|
23 |
+
| [bge-base-en-v1.5-q8_0.gguf](https://huggingface.co/CompendiumLabs/bge-base-en-v1.5-gguf/blob/main/bge-base-en-v1.5-q8_0.gguf) | Q8_0 | 118 MB |
|
24 |
+
| [bge-base-en-v1.5-q4_k_m.gguf](https://huggingface.co/CompendiumLabs/bge-base-en-v1.5-gguf/blob/main/bge-base-en-v1.5-q4_k_m.gguf) | Q4_K_M | 68 MB |
|
25 |
+
|
26 |
+
</div>
|
27 |
+
|
28 |
+
<br/>
|
29 |
|
30 |
# Inference
|
31 |
|