morriszms commited on
Commit
519cb2f
1 Parent(s): 884d4e4

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,16 +1,17 @@
1
  ---
2
- language:
3
- - en
4
- library_name: transformers
5
  license: gemma
 
 
 
 
 
 
 
6
  tags:
7
- - unsloth
8
- - transformers
9
- - gemma2
10
- - gemma
11
  - TensorBlock
12
  - GGUF
13
- base_model: unsloth/gemma-2-9b-it
14
  ---
15
 
16
  <div style="width: auto; margin-left: auto; margin-right: auto">
@@ -24,13 +25,12 @@ base_model: unsloth/gemma-2-9b-it
24
  </div>
25
  </div>
26
 
27
- ## unsloth/gemma-2-9b-it - GGUF
28
 
29
- This repo contains GGUF format model files for [unsloth/gemma-2-9b-it](https://huggingface.co/unsloth/gemma-2-9b-it).
30
 
31
  The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
32
 
33
-
34
  <div style="text-align: left; margin: 20px 0;">
35
  <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
36
  Run them on the TensorBlock client using your local machine ↗
@@ -39,7 +39,6 @@ The files were quantized using machines provided by [TensorBlock](https://tensor
39
 
40
  ## Prompt template
41
 
42
-
43
  ```
44
  <bos><start_of_turn>user
45
  {prompt}<end_of_turn>
@@ -50,18 +49,18 @@ The files were quantized using machines provided by [TensorBlock](https://tensor
50
 
51
  | Filename | Quant type | File Size | Description |
52
  | -------- | ---------- | --------- | ----------- |
53
- | [gemma-2-9b-it-Q2_K.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q2_K.gguf) | Q2_K | 3.544 GB | smallest, significant quality loss - not recommended for most purposes |
54
- | [gemma-2-9b-it-Q3_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q3_K_S.gguf) | Q3_K_S | 4.040 GB | very small, high quality loss |
55
- | [gemma-2-9b-it-Q3_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q3_K_M.gguf) | Q3_K_M | 4.435 GB | very small, high quality loss |
56
- | [gemma-2-9b-it-Q3_K_L.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q3_K_L.gguf) | Q3_K_L | 4.780 GB | small, substantial quality loss |
57
- | [gemma-2-9b-it-Q4_0.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q4_0.gguf) | Q4_0 | 5.069 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
58
- | [gemma-2-9b-it-Q4_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q4_K_S.gguf) | Q4_K_S | 5.103 GB | small, greater quality loss |
59
- | [gemma-2-9b-it-Q4_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q4_K_M.gguf) | Q4_K_M | 5.365 GB | medium, balanced quality - recommended |
60
- | [gemma-2-9b-it-Q5_0.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q5_0.gguf) | Q5_0 | 6.038 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
61
- | [gemma-2-9b-it-Q5_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q5_K_S.gguf) | Q5_K_S | 6.038 GB | large, low quality loss - recommended |
62
- | [gemma-2-9b-it-Q5_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q5_K_M.gguf) | Q5_K_M | 6.191 GB | large, very low quality loss - recommended |
63
- | [gemma-2-9b-it-Q6_K.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q6_K.gguf) | Q6_K | 7.068 GB | very large, extremely low quality loss |
64
- | [gemma-2-9b-it-Q8_0.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q8_0.gguf) | Q8_0 | 9.152 GB | very large, extremely low quality loss - not recommended |
65
 
66
 
67
  ## Downloading instruction
 
1
  ---
 
 
 
2
  license: gemma
3
+ library_name: transformers
4
+ pipeline_tag: text-generation
5
+ extra_gated_heading: Access Gemma on Hugging Face
6
+ extra_gated_prompt: To access Gemma on Hugging Face, you’re required to review and
7
+ agree to Google’s usage license. To do this, please ensure you’re logged in to Hugging
8
+ Face and click below. Requests are processed immediately.
9
+ extra_gated_button_content: Acknowledge license
10
  tags:
11
+ - conversational
 
 
 
12
  - TensorBlock
13
  - GGUF
14
+ base_model: google/gemma-2-9b-it
15
  ---
16
 
17
  <div style="width: auto; margin-left: auto; margin-right: auto">
 
25
  </div>
26
  </div>
27
 
28
+ ## google/gemma-2-9b-it - GGUF
29
 
30
+ This repo contains GGUF format model files for [google/gemma-2-9b-it](https://huggingface.co/google/gemma-2-9b-it).
31
 
32
  The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
33
 
 
34
  <div style="text-align: left; margin: 20px 0;">
35
  <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
36
  Run them on the TensorBlock client using your local machine ↗
 
39
 
40
  ## Prompt template
41
 
 
42
  ```
43
  <bos><start_of_turn>user
44
  {prompt}<end_of_turn>
 
49
 
50
  | Filename | Quant type | File Size | Description |
51
  | -------- | ---------- | --------- | ----------- |
52
+ | [gemma-2-9b-it-Q2_K.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q2_K.gguf) | Q2_K | 3.805 GB | smallest, significant quality loss - not recommended for most purposes |
53
+ | [gemma-2-9b-it-Q3_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q3_K_S.gguf) | Q3_K_S | 4.338 GB | very small, high quality loss |
54
+ | [gemma-2-9b-it-Q3_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q3_K_M.gguf) | Q3_K_M | 4.762 GB | very small, high quality loss |
55
+ | [gemma-2-9b-it-Q3_K_L.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q3_K_L.gguf) | Q3_K_L | 5.132 GB | small, substantial quality loss |
56
+ | [gemma-2-9b-it-Q4_0.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q4_0.gguf) | Q4_0 | 5.443 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
57
+ | [gemma-2-9b-it-Q4_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q4_K_S.gguf) | Q4_K_S | 5.479 GB | small, greater quality loss |
58
+ | [gemma-2-9b-it-Q4_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q4_K_M.gguf) | Q4_K_M | 5.761 GB | medium, balanced quality - recommended |
59
+ | [gemma-2-9b-it-Q5_0.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q5_0.gguf) | Q5_0 | 6.484 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
60
+ | [gemma-2-9b-it-Q5_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q5_K_S.gguf) | Q5_K_S | 6.484 GB | large, low quality loss - recommended |
61
+ | [gemma-2-9b-it-Q5_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q5_K_M.gguf) | Q5_K_M | 6.647 GB | large, very low quality loss - recommended |
62
+ | [gemma-2-9b-it-Q6_K.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q6_K.gguf) | Q6_K | 7.589 GB | very large, extremely low quality loss |
63
+ | [gemma-2-9b-it-Q8_0.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-GGUF/blob/main/gemma-2-9b-it-Q8_0.gguf) | Q8_0 | 9.827 GB | very large, extremely low quality loss - not recommended |
64
 
65
 
66
  ## Downloading instruction
gemma-2-9b-it-Q2_K.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:82250c200ce7a505ea787adc6a4ef1113de77ceff6eec6c9d6423bb46c7968ec
3
- size 3805398080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fafd2beba8cd73ff478bdce43db425ab233186b23d1e8398d2857cf183fff21
3
+ size 3805398208
gemma-2-9b-it-Q3_K_L.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:22a463c6e069aa07279b3d90b2ed23d3ad38fb8f28559bd32209a29bcb8c46a7
3
- size 5132452928
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60f269a9e2d7e347865d037b12f6f15c283c5744c8e2fd169cdc7b32f7e2c953
3
+ size 5132453056
gemma-2-9b-it-Q3_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e034fd6e780ceb3572fb9c088a1c4496731e799cc73da091cf77bea81894ae4
3
- size 4761781312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e7f54dd893bdc59444fd2f226cd1df775a3499b4f28abe995afbb5e31bfc946
3
+ size 4761781440
gemma-2-9b-it-Q3_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:330400b1a66148b2171a00c65bf0f2a1290466d7c1c7ffa2c77023fd1e4c31d1
3
- size 4337665088
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12147f4e4432931ddb6e80141a781a0c2feebf9ba5d76dcd00705df7dd835eb5
3
+ size 4337665216
gemma-2-9b-it-Q4_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:767e3b0c016381522f514efd87e7ea19967ba1734e6dd1aebd7ff1fb1ca073e6
3
- size 5443142720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10b6b9a86d7210121a44c71dcc1f437c5328031c718987442856efe55e7fb855
3
+ size 5443142848
gemma-2-9b-it-Q4_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7e77358a5eb4a49cdbed91502866816c81ba5fa82695c92bed2d18fe24acc133
3
- size 5761057856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef06a5fffb98a53f8e852ff387f81bf8f722b7ea72a07303be166bfe9ddf12e5
3
+ size 5761057984
gemma-2-9b-it-Q4_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:433c90f8b7ac23a3c89f90fc2f9a0bddeede6672106ed7dcf5ba72ff4387cc6e
3
- size 5478925376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b95531dc23ae68ead546d3349e14d789a09f8bb28189cd03ddf6741f664e9785
3
+ size 5478925504
gemma-2-9b-it-Q5_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f76243c22c1230bc5fe99859f3f74b5aca37fb1c7b0749bd9fae4507e3e1b5bd
3
- size 6483592256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c20f8e76334b96b5638a87e8ab9cbbf6ce135068873df2a0c5a2a9070e0ee567
3
+ size 6483592384
gemma-2-9b-it-Q5_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:971890b552b4fc92b0a4a5e7ab980618e93824762e8aca1e974965f674d17436
3
- size 6647366720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ad66079ed811a18147b2ccc4811ea8fc4488c6441358bdac44c1b1e1b0a37b1
3
+ size 6647366848
gemma-2-9b-it-Q5_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a3df6479f610687903338b8030980fb04ff0c22b67869dbfcc6c3c519f80363c
3
- size 6483592256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d31355fef4dc8a6ad2213620c9835486e0399303283e5b7a755366c42eb49be4
3
+ size 6483592384
gemma-2-9b-it-Q6_K.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b5dedc0e8530444718a807007913241c34bb15ea601a60cced655215e27b54f7
3
- size 7589069888
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c85b7aedd545a1c52e87de41d1dbdaf2895c521d77dc6ea2f9f9699a159d03e4
3
+ size 7589070016
gemma-2-9b-it-Q8_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a12f533529fbba551a26a4000ecdcec8db9f141d6d2f8dfd9563b244eacbe377
3
- size 9827148864
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1db97fd35ce33f0746350d7af15ca409c9ffddf524c4db0322e5c2f79c3d5f1c
3
+ size 9827148992