Text Generation
Transformers
GGUF
ibm-granite-code
code
granite
Eval Results
AndreyBest commited on
Commit
9b489d4
1 Parent(s): a280ac9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +263 -3
README.md CHANGED
@@ -1,3 +1,263 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: text-generation
3
+ base_model: ibm-granite/granite-8b-code-base
4
+ inference: false
5
+ license: apache-2.0
6
+ datasets:
7
+ - bigcode/commitpackft
8
+ - TIGER-Lab/MathInstruct
9
+ - meta-math/MetaMathQA
10
+ - glaiveai/glaive-code-assistant-v3
11
+ - glaive-function-calling-v2
12
+ - bugdaryan/sql-create-context-instruction
13
+ - garage-bAInd/Open-Platypus
14
+ - nvidia/HelpSteer
15
+ metrics:
16
+ - code_eval
17
+ library_name: transformers
18
+ tags:
19
+ - code
20
+ - granite
21
+ model-index:
22
+ - name: granite-8b-code-instruct
23
+ results:
24
+ - task:
25
+ type: text-generation
26
+ dataset:
27
+ type: bigcode/humanevalpack
28
+ name: HumanEvalSynthesis(Python)
29
+ metrics:
30
+ - name: pass@1
31
+ type: pass@1
32
+ value: 57.9
33
+ veriefied: false
34
+ - task:
35
+ type: text-generation
36
+ dataset:
37
+ type: bigcode/humanevalpack
38
+ name: HumanEvalSynthesis(JavaScript)
39
+ metrics:
40
+ - name: pass@1
41
+ type: pass@1
42
+ value: 52.4
43
+ veriefied: false
44
+ - task:
45
+ type: text-generation
46
+ dataset:
47
+ type: bigcode/humanevalpack
48
+ name: HumanEvalSynthesis(Java)
49
+ metrics:
50
+ - name: pass@1
51
+ type: pass@1
52
+ value: 58.5
53
+ veriefied: false
54
+ - task:
55
+ type: text-generation
56
+ dataset:
57
+ type: bigcode/humanevalpack
58
+ name: HumanEvalSynthesis(Go)
59
+ metrics:
60
+ - name: pass@1
61
+ type: pass@1
62
+ value: 43.3
63
+ veriefied: false
64
+ - task:
65
+ type: text-generation
66
+ dataset:
67
+ type: bigcode/humanevalpack
68
+ name: HumanEvalSynthesis(C++)
69
+ metrics:
70
+ - name: pass@1
71
+ type: pass@1
72
+ value: 48.2
73
+ veriefied: false
74
+ - task:
75
+ type: text-generation
76
+ dataset:
77
+ type: bigcode/humanevalpack
78
+ name: HumanEvalSynthesis(Rust)
79
+ metrics:
80
+ - name: pass@1
81
+ type: pass@1
82
+ value: 37.2
83
+ veriefied: false
84
+ - task:
85
+ type: text-generation
86
+ dataset:
87
+ type: bigcode/humanevalpack
88
+ name: HumanEvalExplain(Python)
89
+ metrics:
90
+ - name: pass@1
91
+ type: pass@1
92
+ value: 53.0
93
+ veriefied: false
94
+ - task:
95
+ type: text-generation
96
+ dataset:
97
+ type: bigcode/humanevalpack
98
+ name: HumanEvalExplain(JavaScript)
99
+ metrics:
100
+ - name: pass@1
101
+ type: pass@1
102
+ value: 42.7
103
+ veriefied: false
104
+ - task:
105
+ type: text-generation
106
+ dataset:
107
+ type: bigcode/humanevalpack
108
+ name: HumanEvalExplain(Java)
109
+ metrics:
110
+ - name: pass@1
111
+ type: pass@1
112
+ value: 52.4
113
+ veriefied: false
114
+ - task:
115
+ type: text-generation
116
+ dataset:
117
+ type: bigcode/humanevalpack
118
+ name: HumanEvalExplain(Go)
119
+ metrics:
120
+ - name: pass@1
121
+ type: pass@1
122
+ value: 36.6
123
+ veriefied: false
124
+ - task:
125
+ type: text-generation
126
+ dataset:
127
+ type: bigcode/humanevalpack
128
+ name: HumanEvalExplain(C++)
129
+ metrics:
130
+ - name: pass@1
131
+ type: pass@1
132
+ value: 43.9
133
+ veriefied: false
134
+ - task:
135
+ type: text-generation
136
+ dataset:
137
+ type: bigcode/humanevalpack
138
+ name: HumanEvalExplain(Rust)
139
+ metrics:
140
+ - name: pass@1
141
+ type: pass@1
142
+ value: 16.5
143
+ veriefied: false
144
+ - task:
145
+ type: text-generation
146
+ dataset:
147
+ type: bigcode/humanevalpack
148
+ name: HumanEvalFix(Python)
149
+ metrics:
150
+ - name: pass@1
151
+ type: pass@1
152
+ value: 39.6
153
+ veriefied: false
154
+ - task:
155
+ type: text-generation
156
+ dataset:
157
+ type: bigcode/humanevalpack
158
+ name: HumanEvalFix(JavaScript)
159
+ metrics:
160
+ - name: pass@1
161
+ type: pass@1
162
+ value: 40.9
163
+ veriefied: false
164
+ - task:
165
+ type: text-generation
166
+ dataset:
167
+ type: bigcode/humanevalpack
168
+ name: HumanEvalFix(Java)
169
+ metrics:
170
+ - name: pass@1
171
+ type: pass@1
172
+ value: 48.2
173
+ veriefied: false
174
+ - task:
175
+ type: text-generation
176
+ dataset:
177
+ type: bigcode/humanevalpack
178
+ name: HumanEvalFix(Go)
179
+ metrics:
180
+ - name: pass@1
181
+ type: pass@1
182
+ value: 41.5
183
+ veriefied: false
184
+ - task:
185
+ type: text-generation
186
+ dataset:
187
+ type: bigcode/humanevalpack
188
+ name: HumanEvalFix(C++)
189
+ metrics:
190
+ - name: pass@1
191
+ type: pass@1
192
+ value: 39.0
193
+ veriefied: false
194
+ - task:
195
+ type: text-generation
196
+ dataset:
197
+ type: bigcode/humanevalpack
198
+ name: HumanEvalFix(Rust)
199
+ metrics:
200
+ - name: pass@1
201
+ type: pass@1
202
+ value: 32.9
203
+ veriefied: false
204
+ ---
205
+
206
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64a28db2f1968b7d7f357182/rOiYpb6GH0VhWZRmwcOCP.png)
207
+ *This model was quantized by [SanctumAI](https://sanctum.ai). To leave feedback, join our community in [Discord](https://discord.gg/7ZNE78HJKh).*
208
+
209
+ # Granite 8B Code Instruct GGUF
210
+
211
+ **Model creator:** [ibm-granite](https://huggingface.co/ibm-granite)<br>
212
+ **Original model**: [granite-8b-code-instruct](https://huggingface.co/ibm-granite/granite-8b-code-instruct)<br>
213
+
214
+ ## Model Summary:
215
+
216
+ **Granite-8B-Code-Instruct** is a 8B parameter model fine tuned from *Granite-8B-Code-Base* on a combination of **permissively licensed** instruction data to enhance instruction following capabilities including logical reasoning and problem-solving skills.
217
+
218
+ - **Developers:** IBM Research
219
+ - **GitHub Repository:** [ibm-granite/granite-code-models](https://github.com/ibm-granite/granite-code-models)
220
+ - **Paper:** [Granite Code Models: A Family of Open Foundation Models for Code Intelligence](https://arxiv.org/abs/2405.04324)
221
+ - **Release Date**: May 6th, 2024
222
+ - **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0).
223
+
224
+ ## Prompt Template:
225
+
226
+ If you're using Sanctum app, simply use `IBM Granite Code` model preset.
227
+
228
+ Prompt template:
229
+
230
+ ```
231
+ System:
232
+ {system_prompt}
233
+ Question:
234
+ {prompt}
235
+ Answer:
236
+
237
+ ```
238
+
239
+ ## Hardware Requirements Estimate
240
+
241
+ | Name | Quant method | Size | Memory (RAM, vRAM) required |
242
+ | ---- | ---- | ---- | ---- |
243
+ | [granite-8b-code-instruct.Q2_K.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.Q2_K.gguf) | Q2_K | 3.06 GB | 7.47 GB |
244
+ | [granite-8b-code-instruct.Q3_K_S.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.Q3_K_S.gguf) | Q3_K_S | 3.55 GB | ? |
245
+ | [granite-8b-code-instruct.Q3_K_M.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.Q3_K_M.gguf) | Q3_K_M | 3.94 GB | ? |
246
+ | [granite-8b-code-instruct.Q3_K_L.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.Q3_K_L.gguf) | Q3_K_L | 4.29 GB | ? |
247
+ | [granite-8b-code-instruct.Q4_0.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.Q4_0.gguf) | Q4_0 | 4.59 GB | ? |
248
+ | [granite-8b-code-instruct.Q4_K_S.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.Q4_K_S.gguf) | Q4_K_S | 4.62 GB | ? |
249
+ | [granite-8b-code-instruct.Q4_K_M.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.Q4_K_M.gguf) | Q4_K_M | 4.88 GB | ? |
250
+ | [granite-8b-code-instruct.Q4_K.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.Q4_K.gguf) | Q4_K | 4.88 GB | ? |
251
+ | [granite-8b-code-instruct.Q4_1.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.Q4_1.gguf) | Q4_1 | 5.08 GB | ? |
252
+ | [granite-8b-code-instruct.Q5_0.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.Q5_0.gguf) | Q5_0 | 5.57 GB | ? |
253
+ | [granite-8b-code-instruct.Q5_K_S.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.Q5_K_S.gguf) | Q5_K_S | 5.57 GB | ? |
254
+ | [granite-8b-code-instruct.Q5_K_M.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.Q5_K_M.gguf) | Q5_K_M | 5.72 GB | ? |
255
+ | [granite-8b-code-instruct.Q5_K.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.Q5_K.gguf) | Q5_K | 5.72 GB | ? |
256
+ | [granite-8b-code-instruct.Q5_1.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.Q5_1.gguf) | Q5_1 | 6.06 GB | ? |
257
+ | [granite-8b-code-instruct.Q6_K.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.Q6_K.gguf) | Q6_K | 6.62 GB | ? |
258
+ | [granite-8b-code-instruct.Q8_0.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.Q8_0.gguf) | Q8_0 | 8.57 GB | ? |
259
+ | [granite-8b-code-instruct.f16.gguf](https://huggingface.co/SanctumAI/granite-8b-code-instruct-GGUF/blob/main/granite-8b-code-instruct.f16.gguf) | f16 | 16.12 GB | 19.62 GB |
260
+
261
+ ## Disclaimer
262
+
263
+ Sanctum is not the creator, originator, or owner of any Model featured in the Models section of the Sanctum application. Each Model is created and provided by third parties. Sanctum does not endorse, support, represent or guarantee the completeness, truthfulness, accuracy, or reliability of any Model listed there. You understand that supported Models can produce content that might be offensive, harmful, inaccurate or otherwise inappropriate, or deceptive. Each Model is the sole responsibility of the person or entity who originated such Model. Sanctum may not monitor or control the Models supported and cannot, and does not, take responsibility for any such Model. Sanctum disclaims all warranties or guarantees about the accuracy, reliability or benefits of the Models. Sanctum further disclaims any warranty that the Model will meet your requirements, be secure, uninterrupted or available at any time or location, or error-free, viruses-free, or that any errors will be corrected, or otherwise. You will be solely responsible for any damage resulting from your use of or access to the Models, your downloading of any Model, or use of any other Model provided by or through Sanctum.