iamlemec commited on
Commit
c809bd6
1 Parent(s): 0571a2f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -5
README.md CHANGED
@@ -8,11 +8,24 @@ license: mit
8
 
9
  Source model: https://huggingface.co/BAAI/bge-base-en-v1.5
10
 
11
- Quantization types:
12
- - F32 (436 MB)
13
- - F16 (219 MB)
14
- - Q8_0 (118 MB)
15
- - Q4_K_M (68 MB)
 
 
 
 
 
 
 
 
 
 
 
 
 
16
 
17
  # Inference
18
 
 
8
 
9
  Source model: https://huggingface.co/BAAI/bge-base-en-v1.5
10
 
11
+ Quantized and unquantized embedding models for use with `llama.cpp`.
12
+
13
+ <br/>
14
+
15
+ # Files Available
16
+
17
+ <div style="width: 500px; margin: 0;">
18
+
19
+ | Filename | Quantization | Size |
20
+ |:-------- | ------------ | ---- |
21
+ | [bge-base-en-v1.5-f32.gguf](https://huggingface.co/CompendiumLabs/bge-base-en-v1.5-gguf/blob/main/bge-base-en-v1.5-f32.gguf) | F32 | 436 MB |
22
+ | [bge-base-en-v1.5-f16.gguf](https://huggingface.co/CompendiumLabs/bge-base-en-v1.5-gguf/blob/main/bge-base-en-v1.5-f16.gguf) | F16 | 219 MB |
23
+ | [bge-base-en-v1.5-q8_0.gguf](https://huggingface.co/CompendiumLabs/bge-base-en-v1.5-gguf/blob/main/bge-base-en-v1.5-q8_0.gguf) | Q8_0 | 118 MB |
24
+ | [bge-base-en-v1.5-q4_k_m.gguf](https://huggingface.co/CompendiumLabs/bge-base-en-v1.5-gguf/blob/main/bge-base-en-v1.5-q4_k_m.gguf) | Q4_K_M | 68 MB |
25
+
26
+ </div>
27
+
28
+ <br/>
29
 
30
  # Inference
31