Upload folder using huggingface_hub
Browse files- .gitattributes +9 -0
- README.md +51 -0
- model-bf16.gguf +3 -0
- model-f16.gguf +3 -0
- model-f32.gguf +3 -0
- model-q4_k_m.gguf +3 -0
- model-q4_k_s.gguf +3 -0
- model-q5_k_m.gguf +3 -0
- model-q5_k_s.gguf +3 -0
- model-q6_k.gguf +3 -0
- model-q8_0.gguf +3 -0
.gitattributes
CHANGED
@@ -33,3 +33,12 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
model-bf16.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
+
model-f16.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
+
model-f32.gguf filter=lfs diff=lfs merge=lfs -text
|
39 |
+
model-q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
40 |
+
model-q4_k_s.gguf filter=lfs diff=lfs merge=lfs -text
|
41 |
+
model-q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
42 |
+
model-q5_k_s.gguf filter=lfs diff=lfs merge=lfs -text
|
43 |
+
model-q6_k.gguf filter=lfs diff=lfs merge=lfs -text
|
44 |
+
model-q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: cc-by-4.0
|
3 |
+
base_model: hiieu/halong_embedding
|
4 |
+
language: ["vi"]
|
5 |
+
library_name: sentence-transformers
|
6 |
+
pipeline_tag: sentence-similarity
|
7 |
+
inference: false
|
8 |
+
---
|
9 |
+
|
10 |
+
# `hiieu/halong_embedding` in GGUF format
|
11 |
+
|
12 |
+
original: https://huggingface.co/hiieu/halong_embedding
|
13 |
+
|
14 |
+
quantization:
|
15 |
+
```bash
|
16 |
+
REL=b3827 # can change to a later release
|
17 |
+
wget https://github.com/ggerganov/llama.cpp/releases/download/$REL/llama-$REL-bin-ubuntu-x64.zip --content-disposition --continue &> /dev/null
|
18 |
+
wget https://github.com/ggerganov/llama.cpp/archive/refs/tags/$REL.zip --content-disposition --continue &> /dev/null
|
19 |
+
unzip -q llama-$REL-bin-ubuntu-x64.zip
|
20 |
+
unzip -q llama.cpp-$REL.zip
|
21 |
+
mv llama.cpp-$REL/* .
|
22 |
+
rm -r llama.cpp-$REL/ llama-$REL-bin-ubuntu-x64.zip llama.cpp-$REL.zip
|
23 |
+
pip install -q -r requirements.txt
|
24 |
+
|
25 |
+
rm -rf models/tmp/
|
26 |
+
git clone --depth=1 --single-branch https://huggingface.co/hiieu/halong_embedding models/tmp
|
27 |
+
huggingface-cli download intfloat/multilingual-e5-base sentencepiece.bpe.model --local-dir models/tmp
|
28 |
+
python convert_hf_to_gguf.py models/tmp/ --outfile model-f32.gguf --outtype f32
|
29 |
+
|
30 |
+
build/bin/llama-quantize model-f32.gguf model-f16.gguf f16 2> /dev/null
|
31 |
+
build/bin/llama-quantize model-f32.gguf model-bf16.gguf bf16 2> /dev/null
|
32 |
+
build/bin/llama-quantize model-f32.gguf model-q8_0.gguf q8_0 2> /dev/null
|
33 |
+
build/bin/llama-quantize model-f32.gguf model-q6_k.gguf q6_k 2> /dev/null
|
34 |
+
build/bin/llama-quantize model-f32.gguf model-q5_k_m.gguf q5_k_m 2> /dev/null
|
35 |
+
build/bin/llama-quantize model-f32.gguf model-q5_k_s.gguf q5_k_s 2> /dev/null
|
36 |
+
build/bin/llama-quantize model-f32.gguf model-q4_k_m.gguf q4_k_m 2> /dev/null
|
37 |
+
build/bin/llama-quantize model-f32.gguf model-q4_k_s.gguf q4_k_s 2> /dev/null
|
38 |
+
|
39 |
+
rm -rf models/yolo/
|
40 |
+
mkdir -p models/yolo
|
41 |
+
mv model-*.gguf models/yolo/
|
42 |
+
touch models/yolo/README.md
|
43 |
+
huggingface-cli upload halong-embedding-gguf models/yolo .
|
44 |
+
```
|
45 |
+
|
46 |
+
usage:
|
47 |
+
```bash
|
48 |
+
build/bin/llama-embedding -m model-q5_k_m.gguf -p "Cô ấy cười nói suốt cả ngày" --embd-output-format array 2> /dev/null
|
49 |
+
# OR
|
50 |
+
build/bin/llama-server --embedding -c 512 -m model-q5_k_m.gguf
|
51 |
+
```
|
model-bf16.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a4675ee972142e8d697d3252a5e75fd5fcf38eae30702b7edf43c2d8e137e211
|
3 |
+
size 562765792
|
model-f16.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5402c0f0210ddbd6dfe9f8ec48ac079ce6f410501228537370dc81675f8527f2
|
3 |
+
size 562765792
|
model-f32.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fa77076af3abf138c1acdc3229e3f29d93b795e7620508f197254bcbba32e32d
|
3 |
+
size 1116638176
|
model-q4_k_m.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:619c5b87e4ba33fe4a2d862ad341cc956a6b0baf3d0b613b1d239e106b892b10
|
3 |
+
size 218732352
|
model-q4_k_s.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cfde0699ceb1fa4f4a07f9d507b032b02d7b8aab2e960040f3f6332dc03738b0
|
3 |
+
size 214760256
|
model-q5_k_m.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:68171ad40f84b25288b347469147f3e9bf18b5b9d5e0a731b74b0a199deb1fea
|
3 |
+
size 227137344
|
model-q5_k_s.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:59a742f2e225fab6c0fb5f4689d216aa352d69c9bc29a1a6b7d0dbbeb73ccfaa
|
3 |
+
size 224787264
|
model-q6_k.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8772f28a0f5f72254bc1ac3f21f801fa9b032ca405a43b562b10d32dd8bc934c
|
3 |
+
size 236067648
|
model-q8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:91a4d02d544dd90dd04692e73641e89c1a383fce54af813ecdf20124d3973230
|
3 |
+
size 303138112
|