Triangle104 commited on
Commit
c7d76fd
1 Parent(s): 86d98be

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +228 -0
README.md ADDED
@@ -0,0 +1,228 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: llama3.2
5
+ tags:
6
+ - shining-valiant
7
+ - shining-valiant-2
8
+ - valiant
9
+ - valiant-labs
10
+ - llama
11
+ - llama-3.2
12
+ - llama-3.2-instruct
13
+ - llama-3.2-instruct-3b
14
+ - llama-3
15
+ - llama-3-instruct
16
+ - llama-3-instruct-3b
17
+ - 3b
18
+ - science
19
+ - physics
20
+ - biology
21
+ - chemistry
22
+ - compsci
23
+ - computer-science
24
+ - engineering
25
+ - technical
26
+ - conversational
27
+ - chat
28
+ - instruct
29
+ - llama-cpp
30
+ - gguf-my-repo
31
+ base_model: ValiantLabs/Llama3.2-3B-ShiningValiant2
32
+ datasets:
33
+ - sequelbox/Celestia
34
+ - sequelbox/Spurline
35
+ - sequelbox/Supernova
36
+ pipeline_tag: text-generation
37
+ model_type: llama
38
+ model-index:
39
+ - name: Llama3.2-3B-ShiningValiant2
40
+ results:
41
+ - task:
42
+ type: text-generation
43
+ name: Text Generation
44
+ dataset:
45
+ name: Winogrande (5-shot)
46
+ type: winogrande
47
+ args:
48
+ num_few_shot: 5
49
+ metrics:
50
+ - type: acc
51
+ value: 69.14
52
+ name: acc
53
+ - task:
54
+ type: text-generation
55
+ name: Text Generation
56
+ dataset:
57
+ name: MMLU College Biology (5-shot)
58
+ type: mmlu
59
+ args:
60
+ num_few_shot: 5
61
+ metrics:
62
+ - type: acc
63
+ value: 64.58
64
+ name: acc
65
+ - type: acc
66
+ value: 70.32
67
+ name: acc
68
+ - type: acc
69
+ value: 44.0
70
+ name: acc
71
+ - type: acc
72
+ value: 50.25
73
+ name: acc
74
+ - type: acc
75
+ value: 42.16
76
+ name: acc
77
+ - type: acc
78
+ value: 35.76
79
+ name: acc
80
+ - type: acc
81
+ value: 53.19
82
+ name: acc
83
+ - type: acc
84
+ value: 53.0
85
+ name: acc
86
+ - type: acc
87
+ value: 61.0
88
+ name: acc
89
+ - type: acc
90
+ value: 60.53
91
+ name: acc
92
+ - task:
93
+ type: text-generation
94
+ name: Text Generation
95
+ dataset:
96
+ name: IFEval (0-Shot)
97
+ type: HuggingFaceH4/ifeval
98
+ args:
99
+ num_few_shot: 0
100
+ metrics:
101
+ - type: inst_level_strict_acc and prompt_level_strict_acc
102
+ value: 48.9
103
+ name: strict accuracy
104
+ source:
105
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.2-3B-ShiningValiant2
106
+ name: Open LLM Leaderboard
107
+ - task:
108
+ type: text-generation
109
+ name: Text Generation
110
+ dataset:
111
+ name: BBH (3-Shot)
112
+ type: BBH
113
+ args:
114
+ num_few_shot: 3
115
+ metrics:
116
+ - type: acc_norm
117
+ value: 19.11
118
+ name: normalized accuracy
119
+ source:
120
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.2-3B-ShiningValiant2
121
+ name: Open LLM Leaderboard
122
+ - task:
123
+ type: text-generation
124
+ name: Text Generation
125
+ dataset:
126
+ name: MATH Lvl 5 (4-Shot)
127
+ type: hendrycks/competition_math
128
+ args:
129
+ num_few_shot: 4
130
+ metrics:
131
+ - type: exact_match
132
+ value: 9.14
133
+ name: exact match
134
+ source:
135
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.2-3B-ShiningValiant2
136
+ name: Open LLM Leaderboard
137
+ - task:
138
+ type: text-generation
139
+ name: Text Generation
140
+ dataset:
141
+ name: GPQA (0-shot)
142
+ type: Idavidrein/gpqa
143
+ args:
144
+ num_few_shot: 0
145
+ metrics:
146
+ - type: acc_norm
147
+ value: 3.02
148
+ name: acc_norm
149
+ source:
150
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.2-3B-ShiningValiant2
151
+ name: Open LLM Leaderboard
152
+ - task:
153
+ type: text-generation
154
+ name: Text Generation
155
+ dataset:
156
+ name: MuSR (0-shot)
157
+ type: TAUR-Lab/MuSR
158
+ args:
159
+ num_few_shot: 0
160
+ metrics:
161
+ - type: acc_norm
162
+ value: 5.49
163
+ name: acc_norm
164
+ source:
165
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.2-3B-ShiningValiant2
166
+ name: Open LLM Leaderboard
167
+ - task:
168
+ type: text-generation
169
+ name: Text Generation
170
+ dataset:
171
+ name: MMLU-PRO (5-shot)
172
+ type: TIGER-Lab/MMLU-Pro
173
+ config: main
174
+ split: test
175
+ args:
176
+ num_few_shot: 5
177
+ metrics:
178
+ - type: acc
179
+ value: 19.1
180
+ name: accuracy
181
+ source:
182
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.2-3B-ShiningValiant2
183
+ name: Open LLM Leaderboard
184
+ ---
185
+
186
+ # Triangle104/Llama3.2-3B-ShiningValiant2-Q8_0-GGUF
187
+ This model was converted to GGUF format from [`ValiantLabs/Llama3.2-3B-ShiningValiant2`](https://huggingface.co/ValiantLabs/Llama3.2-3B-ShiningValiant2) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
188
+ Refer to the [original model card](https://huggingface.co/ValiantLabs/Llama3.2-3B-ShiningValiant2) for more details on the model.
189
+
190
+ ## Use with llama.cpp
191
+ Install llama.cpp through brew (works on Mac and Linux)
192
+
193
+ ```bash
194
+ brew install llama.cpp
195
+
196
+ ```
197
+ Invoke the llama.cpp server or the CLI.
198
+
199
+ ### CLI:
200
+ ```bash
201
+ llama-cli --hf-repo Triangle104/Llama3.2-3B-ShiningValiant2-Q8_0-GGUF --hf-file llama3.2-3b-shiningvaliant2-q8_0.gguf -p "The meaning to life and the universe is"
202
+ ```
203
+
204
+ ### Server:
205
+ ```bash
206
+ llama-server --hf-repo Triangle104/Llama3.2-3B-ShiningValiant2-Q8_0-GGUF --hf-file llama3.2-3b-shiningvaliant2-q8_0.gguf -c 2048
207
+ ```
208
+
209
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
210
+
211
+ Step 1: Clone llama.cpp from GitHub.
212
+ ```
213
+ git clone https://github.com/ggerganov/llama.cpp
214
+ ```
215
+
216
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
217
+ ```
218
+ cd llama.cpp && LLAMA_CURL=1 make
219
+ ```
220
+
221
+ Step 3: Run inference through the main binary.
222
+ ```
223
+ ./llama-cli --hf-repo Triangle104/Llama3.2-3B-ShiningValiant2-Q8_0-GGUF --hf-file llama3.2-3b-shiningvaliant2-q8_0.gguf -p "The meaning to life and the universe is"
224
+ ```
225
+ or
226
+ ```
227
+ ./llama-server --hf-repo Triangle104/Llama3.2-3B-ShiningValiant2-Q8_0-GGUF --hf-file llama3.2-3b-shiningvaliant2-q8_0.gguf -c 2048
228
+ ```