morriszms commited on
Commit
5cb0ef4
·
verified ·
1 Parent(s): 063aa0e

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,16 +1,14 @@
1
  ---
2
- language:
3
- - en
4
  library_name: transformers
5
- license: apache-2.0
6
  tags:
7
- - unsloth
8
- - transformers
9
- - gemma
10
- - gemma-2b
11
  - TensorBlock
12
  - GGUF
13
- base_model: unsloth/gemma-2b
 
 
 
 
 
14
  ---
15
 
16
  <div style="width: auto; margin-left: auto; margin-right: auto">
@@ -24,13 +22,12 @@ base_model: unsloth/gemma-2b
24
  </div>
25
  </div>
26
 
27
- ## unsloth/gemma-2b - GGUF
28
 
29
- This repo contains GGUF format model files for [unsloth/gemma-2b](https://huggingface.co/unsloth/gemma-2b).
30
 
31
  The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
32
 
33
-
34
  <div style="text-align: left; margin: 20px 0;">
35
  <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
36
  Run them on the TensorBlock client using your local machine ↗
@@ -39,7 +36,6 @@ The files were quantized using machines provided by [TensorBlock](https://tensor
39
 
40
  ## Prompt template
41
 
42
-
43
  ```
44
 
45
  ```
@@ -48,18 +44,18 @@ The files were quantized using machines provided by [TensorBlock](https://tensor
48
 
49
  | Filename | Quant type | File Size | Description |
50
  | -------- | ---------- | --------- | ----------- |
51
- | [gemma-2b-Q2_K.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q2_K.gguf) | Q2_K | 1.078 GB | smallest, significant quality loss - not recommended for most purposes |
52
- | [gemma-2b-Q3_K_S.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q3_K_S.gguf) | Q3_K_S | 1.200 GB | very small, high quality loss |
53
- | [gemma-2b-Q3_K_M.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q3_K_M.gguf) | Q3_K_M | 1.289 GB | very small, high quality loss |
54
- | [gemma-2b-Q3_K_L.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q3_K_L.gguf) | Q3_K_L | 1.365 GB | small, substantial quality loss |
55
- | [gemma-2b-Q4_0.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q4_0.gguf) | Q4_0 | 1.445 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
56
- | [gemma-2b-Q4_K_S.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q4_K_S.gguf) | Q4_K_S | 1.453 GB | small, greater quality loss |
57
- | [gemma-2b-Q4_K_M.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q4_K_M.gguf) | Q4_K_M | 1.518 GB | medium, balanced quality - recommended |
58
- | [gemma-2b-Q5_0.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q5_0.gguf) | Q5_0 | 1.675 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
59
- | [gemma-2b-Q5_K_S.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q5_K_S.gguf) | Q5_K_S | 1.675 GB | large, low quality loss - recommended |
60
- | [gemma-2b-Q5_K_M.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q5_K_M.gguf) | Q5_K_M | 1.713 GB | large, very low quality loss - recommended |
61
- | [gemma-2b-Q6_K.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q6_K.gguf) | Q6_K | 1.921 GB | very large, extremely low quality loss |
62
- | [gemma-2b-Q8_0.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q8_0.gguf) | Q8_0 | 2.486 GB | very large, extremely low quality loss - not recommended |
63
 
64
 
65
  ## Downloading instruction
 
1
  ---
 
 
2
  library_name: transformers
 
3
  tags:
 
 
 
 
4
  - TensorBlock
5
  - GGUF
6
+ extra_gated_heading: Access Gemma on Hugging Face
7
+ extra_gated_prompt: To access Gemma on Hugging Face, you’re required to review and
8
+ agree to Google’s usage license. To do this, please ensure you’re logged-in to Hugging
9
+ Face and click below. Requests are processed immediately.
10
+ extra_gated_button_content: Acknowledge license
11
+ base_model: 4bit/gemma-2b
12
  ---
13
 
14
  <div style="width: auto; margin-left: auto; margin-right: auto">
 
22
  </div>
23
  </div>
24
 
25
+ ## 4bit/gemma-2b - GGUF
26
 
27
+ This repo contains GGUF format model files for [4bit/gemma-2b](https://huggingface.co/4bit/gemma-2b).
28
 
29
  The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
30
 
 
31
  <div style="text-align: left; margin: 20px 0;">
32
  <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
33
  Run them on the TensorBlock client using your local machine ↗
 
36
 
37
  ## Prompt template
38
 
 
39
  ```
40
 
41
  ```
 
44
 
45
  | Filename | Quant type | File Size | Description |
46
  | -------- | ---------- | --------- | ----------- |
47
+ | [gemma-2b-Q2_K.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q2_K.gguf) | Q2_K | 1.158 GB | smallest, significant quality loss - not recommended for most purposes |
48
+ | [gemma-2b-Q3_K_S.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q3_K_S.gguf) | Q3_K_S | 1.288 GB | very small, high quality loss |
49
+ | [gemma-2b-Q3_K_M.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q3_K_M.gguf) | Q3_K_M | 1.384 GB | very small, high quality loss |
50
+ | [gemma-2b-Q3_K_L.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q3_K_L.gguf) | Q3_K_L | 1.466 GB | small, substantial quality loss |
51
+ | [gemma-2b-Q4_0.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q4_0.gguf) | Q4_0 | 1.551 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
52
+ | [gemma-2b-Q4_K_S.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q4_K_S.gguf) | Q4_K_S | 1.560 GB | small, greater quality loss |
53
+ | [gemma-2b-Q4_K_M.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q4_K_M.gguf) | Q4_K_M | 1.630 GB | medium, balanced quality - recommended |
54
+ | [gemma-2b-Q5_0.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q5_0.gguf) | Q5_0 | 1.799 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
55
+ | [gemma-2b-Q5_K_S.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q5_K_S.gguf) | Q5_K_S | 1.799 GB | large, low quality loss - recommended |
56
+ | [gemma-2b-Q5_K_M.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q5_K_M.gguf) | Q5_K_M | 1.840 GB | large, very low quality loss - recommended |
57
+ | [gemma-2b-Q6_K.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q6_K.gguf) | Q6_K | 2.062 GB | very large, extremely low quality loss |
58
+ | [gemma-2b-Q8_0.gguf](https://huggingface.co/tensorblock/gemma-2b-GGUF/blob/main/gemma-2b-Q8_0.gguf) | Q8_0 | 2.669 GB | very large, extremely low quality loss - not recommended |
59
 
60
 
61
  ## Downloading instruction
gemma-2b-Q2_K.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:82ccf555f4b1756c88b18ca31561d32f888233aba98d44b2d013832d050c6e12
3
- size 1157923456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e1cee3d7cf3521e60bf37967f4f1fbaee4d4d876eae9cf50406fc99b688b521
3
+ size 1157924224
gemma-2b-Q3_K_L.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ca2ee748240477b2573ae45f2d3bea954c2be4241c2b33609cff4a49f9d5f6c
3
- size 1465590400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cb2887d2243c782f9692318383808749c36291d95dd9a1d2cbc5968a8a99ae7
3
+ size 1465591168
gemma-2b-Q3_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:208f1903873fc30e05d1c5e78733e877674003eb48ba3467f005a4c28b8a3abd
3
- size 1383801472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a280d6a12343c8543819a0f5f5b92475c322b0f9e53c0cf8f4d340c7f5378fb
3
+ size 1383802240
gemma-2b-Q3_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f74e08a728c37fa76afac7a95ce06a5eb62d8a56c7ea8478d03486784e3b5b74
3
- size 1287979648
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d7fe247a3db2d1750507240b63acee138d62e92b6f81912506b2e8fb61f1a98
3
+ size 1287980416
gemma-2b-Q4_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd346e76ebb5fa902d2d17a8de1bfa266101ff858a9a3c3bb949d8eba8ddbe2d
3
- size 1551188608
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:596c7dd5cb82bf57d6a5d8d58aa65a7cac998f8c27718c96490a2e47e52d4b9d
3
+ size 1551189376
gemma-2b-Q4_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3193f11630610aa7670d1ecf20709f38450a1a247336b06a84df26c8f5ecf9dd
3
- size 1630261888
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba83345dd2428be7a7efd3ced56588bd36dc928e62ad482f8ac8afbe93349226
3
+ size 1630262656
gemma-2b-Q4_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2e7aa81b0eb1394f157af144a489d2ca0cf03713bf92397b6b8d1a6ec26f6722
3
- size 1559839360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d47e3ac5127e534590f79a37fe01c697988ce6dea84b78ba0ef5274a0f14244c
3
+ size 1559840128
gemma-2b-Q5_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b348f50835c6104a47dc83795ce0ea8641b8308773370cb56c6473606e4476ff
3
- size 1798914688
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a46e548541ef31d6ffdc90411289135b5daa73253e6d1e3267ccad5006cfa0a3
3
+ size 1798915456
gemma-2b-Q5_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca71cdcec5ca4fd4b5dd222f2eb189b153a5fe1e99cf560d8aae14cd888e5531
3
- size 1839649408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c108d822416151bf829d9b921f7b44295a646e110a5dea8a56e346f749d067b6
3
+ size 1839650176
gemma-2b-Q5_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:98411c555e27e615388878ba7dd29d4a99bf8d3f03dabf055d9551ab25225fc3
3
- size 1798914688
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9741d7ea232ac7df7db7d1495cddb32ecd72dd44748667497d492b7b47c34876
3
+ size 1798915456
gemma-2b-Q6_K.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e2af91f9b9491aeba7bd7abff46bbfd9964f62180a155021712e908337804457
3
- size 2062123648
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73ec9fd6f8a41627bbb10aefd8fa21eac06ffc40653d60bb2d0b54531f5a129b
3
+ size 2062124416
gemma-2b-Q8_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:880ca33c8b833085c38ec6d075431c2ff42f4e661d4f8a347aa56d44543f8162
3
- size 2669068928
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a239df70504543c1feedc8a314ab84366618adc3a62908793cc286d23ab1c2a9
3
+ size 2669069696