itlwas commited on
Commit
738f0ab
1 Parent(s): e1a4800

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +144 -0
README.md ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: text-generation
3
+ inference: false
4
+ license: apache-2.0
5
+ datasets:
6
+ - codeparrot/github-code-clean
7
+ - bigcode/starcoderdata
8
+ - open-web-math/open-web-math
9
+ - math-ai/StackMathQA
10
+ metrics:
11
+ - code_eval
12
+ library_name: transformers
13
+ tags:
14
+ - code
15
+ - granite
16
+ - llama-cpp
17
+ - gguf-my-repo
18
+ base_model: ibm-granite/granite-3b-code-base-2k
19
+ model-index:
20
+ - name: granite-3b-code-base-2k
21
+ results:
22
+ - task:
23
+ type: text-generation
24
+ dataset:
25
+ name: MBPP
26
+ type: mbpp
27
+ metrics:
28
+ - type: pass@1
29
+ value: 36.0
30
+ name: pass@1
31
+ - task:
32
+ type: text-generation
33
+ dataset:
34
+ name: MBPP+
35
+ type: evalplus/mbppplus
36
+ metrics:
37
+ - type: pass@1
38
+ value: 45.1
39
+ name: pass@1
40
+ - task:
41
+ type: text-generation
42
+ dataset:
43
+ name: HumanEvalSynthesis(Python)
44
+ type: bigcode/humanevalpack
45
+ metrics:
46
+ - type: pass@1
47
+ value: 36.6
48
+ name: pass@1
49
+ - type: pass@1
50
+ value: 37.2
51
+ name: pass@1
52
+ - type: pass@1
53
+ value: 40.9
54
+ name: pass@1
55
+ - type: pass@1
56
+ value: 26.2
57
+ name: pass@1
58
+ - type: pass@1
59
+ value: 35.4
60
+ name: pass@1
61
+ - type: pass@1
62
+ value: 22.0
63
+ name: pass@1
64
+ - type: pass@1
65
+ value: 25.0
66
+ name: pass@1
67
+ - type: pass@1
68
+ value: 18.9
69
+ name: pass@1
70
+ - type: pass@1
71
+ value: 29.9
72
+ name: pass@1
73
+ - type: pass@1
74
+ value: 17.1
75
+ name: pass@1
76
+ - type: pass@1
77
+ value: 26.8
78
+ name: pass@1
79
+ - type: pass@1
80
+ value: 14.0
81
+ name: pass@1
82
+ - type: pass@1
83
+ value: 18.3
84
+ name: pass@1
85
+ - type: pass@1
86
+ value: 23.2
87
+ name: pass@1
88
+ - type: pass@1
89
+ value: 29.9
90
+ name: pass@1
91
+ - type: pass@1
92
+ value: 24.4
93
+ name: pass@1
94
+ - type: pass@1
95
+ value: 16.5
96
+ name: pass@1
97
+ - type: pass@1
98
+ value: 3.7
99
+ name: pass@1
100
+ ---
101
+
102
+ # AIronMind/granite-3b-code-base-2k-Q4_K_M-GGUF
103
+ This model was converted to GGUF format from [`ibm-granite/granite-3b-code-base-2k`](https://huggingface.co/ibm-granite/granite-3b-code-base-2k) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
104
+ Refer to the [original model card](https://huggingface.co/ibm-granite/granite-3b-code-base-2k) for more details on the model.
105
+
106
+ ## Use with llama.cpp
107
+ Install llama.cpp through brew (works on Mac and Linux)
108
+
109
+ ```bash
110
+ brew install llama.cpp
111
+
112
+ ```
113
+ Invoke the llama.cpp server or the CLI.
114
+
115
+ ### CLI:
116
+ ```bash
117
+ llama-cli --hf-repo AIronMind/granite-3b-code-base-2k-Q4_K_M-GGUF --hf-file granite-3b-code-base-2k-q4_k_m.gguf -p "The meaning to life and the universe is"
118
+ ```
119
+
120
+ ### Server:
121
+ ```bash
122
+ llama-server --hf-repo AIronMind/granite-3b-code-base-2k-Q4_K_M-GGUF --hf-file granite-3b-code-base-2k-q4_k_m.gguf -c 2048
123
+ ```
124
+
125
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
126
+
127
+ Step 1: Clone llama.cpp from GitHub.
128
+ ```
129
+ git clone https://github.com/ggerganov/llama.cpp
130
+ ```
131
+
132
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
133
+ ```
134
+ cd llama.cpp && LLAMA_CURL=1 make
135
+ ```
136
+
137
+ Step 3: Run inference through the main binary.
138
+ ```
139
+ ./llama-cli --hf-repo AIronMind/granite-3b-code-base-2k-Q4_K_M-GGUF --hf-file granite-3b-code-base-2k-q4_k_m.gguf -p "The meaning to life and the universe is"
140
+ ```
141
+ or
142
+ ```
143
+ ./llama-server --hf-repo AIronMind/granite-3b-code-base-2k-Q4_K_M-GGUF --hf-file granite-3b-code-base-2k-q4_k_m.gguf -c 2048
144
+ ```