morriszms commited on
Commit
2d0e67c
1 Parent(s): f58f31a

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,18 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Llama-3-70B-Orpo-v0.1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Llama-3-70B-Orpo-v0.1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Llama-3-70B-Orpo-v0.1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Llama-3-70B-Orpo-v0.1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Llama-3-70B-Orpo-v0.1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ Llama-3-70B-Orpo-v0.1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ Llama-3-70B-Orpo-v0.1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ Llama-3-70B-Orpo-v0.1-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ Llama-3-70B-Orpo-v0.1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ Llama-3-70B-Orpo-v0.1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ Llama-3-70B-Orpo-v0.1-Q6_K/Llama-3-70B-Orpo-v0.1-Q6_K-00001-of-00002.gguf filter=lfs diff=lfs merge=lfs -text
47
+ Llama-3-70B-Orpo-v0.1-Q6_K/Llama-3-70B-Orpo-v0.1-Q6_K-00002-of-00002.gguf filter=lfs diff=lfs merge=lfs -text
48
+ Llama-3-70B-Orpo-v0.1-Q8_0/Llama-3-70B-Orpo-v0.1-Q8_0-00001-of-00003.gguf filter=lfs diff=lfs merge=lfs -text
49
+ Llama-3-70B-Orpo-v0.1-Q8_0/Llama-3-70B-Orpo-v0.1-Q8_0-00002-of-00003.gguf filter=lfs diff=lfs merge=lfs -text
50
+ Llama-3-70B-Orpo-v0.1-Q8_0/Llama-3-70B-Orpo-v0.1-Q8_0-00003-of-00003.gguf filter=lfs diff=lfs merge=lfs -text
Llama-3-70B-Orpo-v0.1-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac704ed5ddbf2d45b82818715e903fe298c54f05f26720b3db67033d4735d316
3
+ size 26375127744
Llama-3-70B-Orpo-v0.1-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e1864050fa6348a21a5bbf2f5d7bbc41fe4202040c2b9f2f22cba06ca3f0b47
3
+ size 37140613440
Llama-3-70B-Orpo-v0.1-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4c7668aaebdfcd76d6dc589012f649d5f73bf376ef5ebf33f6f7cfa3c69539a
3
+ size 34267515200
Llama-3-70B-Orpo-v0.1-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ad4394879039f5e5cbf265cd1c33543cdd126201febc8599f1499d603412fb2
3
+ size 30912072000
Llama-3-70B-Orpo-v0.1-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:632e22391fbcec3bba22e25a6d5396f35d6679fcf325ba0f9e7e81cd122e5098
3
+ size 39969755584
Llama-3-70B-Orpo-v0.1-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8a83cddb29ef95c5ee29b4aaaf995652e0ef76ca86f06a7d1c5c0fb6a8e1e7a
3
+ size 42520416704
Llama-3-70B-Orpo-v0.1-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:261d68d66e3c36258251125c351ded91de9d000929ea02f66a10f3302c47b82a
3
+ size 40347242944
Llama-3-70B-Orpo-v0.1-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42241138e1c5b641663586a3b170e7f5f5bf0ab5ad49048da3b74d8d241b8edd
3
+ size 48657471936
Llama-3-70B-Orpo-v0.1-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcac3ed5b36bccb9133b8780ad08aa25bb8cf09e1d2f024803bf755ba65d17ec
3
+ size 49949841856
Llama-3-70B-Orpo-v0.1-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c93cd67429b0bfef01e906c79ba62cd87f89424bc0c7b2a03d0e22e9a2766171
3
+ size 48657471936
Llama-3-70B-Orpo-v0.1-Q6_K/Llama-3-70B-Orpo-v0.1-Q6_K-00001-of-00002.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36469ffd678f4cf699be9cce50a82ad6dc72aba94199018661f63bfdff310721
3
+ size 34847497824
Llama-3-70B-Orpo-v0.1-Q6_K/Llama-3-70B-Orpo-v0.1-Q6_K-00002-of-00002.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a14b973d9e7d979a37423b2e0ba7a451dda41bc1c0f78b88967da268de9c7794
3
+ size 23040672928
Llama-3-70B-Orpo-v0.1-Q8_0/Llama-3-70B-Orpo-v0.1-Q8_0-00001-of-00003.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b72cc027393911d36a8c667d38fc8e4c7373eff076a1b0af59b9c3b25c3343ac
3
+ size 34980046208
Llama-3-70B-Orpo-v0.1-Q8_0/Llama-3-70B-Orpo-v0.1-Q8_0-00002-of-00003.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52552a020ec8185c1dee9a8ed06d40728cb9ef77c728e012f1021cf4677f6853
3
+ size 34949976384
Llama-3-70B-Orpo-v0.1-Q8_0/Llama-3-70B-Orpo-v0.1-Q8_0-00003-of-00003.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42914d01bd4ed31b598e9ef7d4821c799faefd3c51569df5ec2e03b3de9ee2ac
3
+ size 5045062560
README.md ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: llama3
5
+ library_name: transformers
6
+ tags:
7
+ - orpo
8
+ - llama 3
9
+ - rlhf
10
+ - sft
11
+ - TensorBlock
12
+ - GGUF
13
+ base_model: dfurman/Llama-3-70B-Orpo-v0.1
14
+ datasets:
15
+ - mlabonne/orpo-dpo-mix-40k
16
+ model-index:
17
+ - name: Llama-3-70B-Orpo-v0.1
18
+ results:
19
+ - task:
20
+ type: text-generation
21
+ name: Text Generation
22
+ dataset:
23
+ name: IFEval (0-Shot)
24
+ type: HuggingFaceH4/ifeval
25
+ args:
26
+ num_few_shot: 0
27
+ metrics:
28
+ - type: inst_level_strict_acc and prompt_level_strict_acc
29
+ value: 20.49
30
+ name: strict accuracy
31
+ source:
32
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=dfurman/Llama-3-70B-Orpo-v0.1
33
+ name: Open LLM Leaderboard
34
+ - task:
35
+ type: text-generation
36
+ name: Text Generation
37
+ dataset:
38
+ name: BBH (3-Shot)
39
+ type: BBH
40
+ args:
41
+ num_few_shot: 3
42
+ metrics:
43
+ - type: acc_norm
44
+ value: 24.09
45
+ name: normalized accuracy
46
+ source:
47
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=dfurman/Llama-3-70B-Orpo-v0.1
48
+ name: Open LLM Leaderboard
49
+ - task:
50
+ type: text-generation
51
+ name: Text Generation
52
+ dataset:
53
+ name: MATH Lvl 5 (4-Shot)
54
+ type: hendrycks/competition_math
55
+ args:
56
+ num_few_shot: 4
57
+ metrics:
58
+ - type: exact_match
59
+ value: 13.52
60
+ name: exact match
61
+ source:
62
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=dfurman/Llama-3-70B-Orpo-v0.1
63
+ name: Open LLM Leaderboard
64
+ - task:
65
+ type: text-generation
66
+ name: Text Generation
67
+ dataset:
68
+ name: GPQA (0-shot)
69
+ type: Idavidrein/gpqa
70
+ args:
71
+ num_few_shot: 0
72
+ metrics:
73
+ - type: acc_norm
74
+ value: 1.01
75
+ name: acc_norm
76
+ source:
77
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=dfurman/Llama-3-70B-Orpo-v0.1
78
+ name: Open LLM Leaderboard
79
+ - task:
80
+ type: text-generation
81
+ name: Text Generation
82
+ dataset:
83
+ name: MuSR (0-shot)
84
+ type: TAUR-Lab/MuSR
85
+ args:
86
+ num_few_shot: 0
87
+ metrics:
88
+ - type: acc_norm
89
+ value: 16.28
90
+ name: acc_norm
91
+ source:
92
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=dfurman/Llama-3-70B-Orpo-v0.1
93
+ name: Open LLM Leaderboard
94
+ - task:
95
+ type: text-generation
96
+ name: Text Generation
97
+ dataset:
98
+ name: MMLU-PRO (5-shot)
99
+ type: TIGER-Lab/MMLU-Pro
100
+ config: main
101
+ split: test
102
+ args:
103
+ num_few_shot: 5
104
+ metrics:
105
+ - type: acc
106
+ value: 32.14
107
+ name: accuracy
108
+ source:
109
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=dfurman/Llama-3-70B-Orpo-v0.1
110
+ name: Open LLM Leaderboard
111
+ ---
112
+
113
+ <div style="width: auto; margin-left: auto; margin-right: auto">
114
+ <img src="https://i.imgur.com/jC7kdl8.jpeg" alt="TensorBlock" style="width: 100%; min-width: 400px; display: block; margin: auto;">
115
+ </div>
116
+ <div style="display: flex; justify-content: space-between; width: 100%;">
117
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
118
+ <p style="margin-top: 0.5em; margin-bottom: 0em;">
119
+ Feedback and support: TensorBlock's <a href="https://x.com/tensorblock_aoi">Twitter/X</a>, <a href="https://t.me/TensorBlock">Telegram Group</a> and <a href="https://x.com/tensorblock_aoi">Discord server</a>
120
+ </p>
121
+ </div>
122
+ </div>
123
+
124
+ ## dfurman/Llama-3-70B-Orpo-v0.1 - GGUF
125
+
126
+ This repo contains GGUF format model files for [dfurman/Llama-3-70B-Orpo-v0.1](https://huggingface.co/dfurman/Llama-3-70B-Orpo-v0.1).
127
+
128
+ The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4242](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
129
+
130
+ <div style="text-align: left; margin: 20px 0;">
131
+ <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
132
+ Run them on the TensorBlock client using your local machine ↗
133
+ </a>
134
+ </div>
135
+
136
+ ## Prompt template
137
+
138
+ ```
139
+ <|im_start|>system
140
+ {system_prompt}<|im_end|>
141
+ <|im_start|>user
142
+ {prompt}<|im_end|>
143
+ <|im_start|>assistant
144
+ ```
145
+
146
+ ## Model file specification
147
+
148
+ | Filename | Quant type | File Size | Description |
149
+ | -------- | ---------- | --------- | ----------- |
150
+ | [Llama-3-70B-Orpo-v0.1-Q2_K.gguf](https://huggingface.co/tensorblock/Llama-3-70B-Orpo-v0.1-GGUF/blob/main/Llama-3-70B-Orpo-v0.1-Q2_K.gguf) | Q2_K | 26.375 GB | smallest, significant quality loss - not recommended for most purposes |
151
+ | [Llama-3-70B-Orpo-v0.1-Q3_K_S.gguf](https://huggingface.co/tensorblock/Llama-3-70B-Orpo-v0.1-GGUF/blob/main/Llama-3-70B-Orpo-v0.1-Q3_K_S.gguf) | Q3_K_S | 30.912 GB | very small, high quality loss |
152
+ | [Llama-3-70B-Orpo-v0.1-Q3_K_M.gguf](https://huggingface.co/tensorblock/Llama-3-70B-Orpo-v0.1-GGUF/blob/main/Llama-3-70B-Orpo-v0.1-Q3_K_M.gguf) | Q3_K_M | 34.268 GB | very small, high quality loss |
153
+ | [Llama-3-70B-Orpo-v0.1-Q3_K_L.gguf](https://huggingface.co/tensorblock/Llama-3-70B-Orpo-v0.1-GGUF/blob/main/Llama-3-70B-Orpo-v0.1-Q3_K_L.gguf) | Q3_K_L | 37.141 GB | small, substantial quality loss |
154
+ | [Llama-3-70B-Orpo-v0.1-Q4_0.gguf](https://huggingface.co/tensorblock/Llama-3-70B-Orpo-v0.1-GGUF/blob/main/Llama-3-70B-Orpo-v0.1-Q4_0.gguf) | Q4_0 | 39.970 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
155
+ | [Llama-3-70B-Orpo-v0.1-Q4_K_S.gguf](https://huggingface.co/tensorblock/Llama-3-70B-Orpo-v0.1-GGUF/blob/main/Llama-3-70B-Orpo-v0.1-Q4_K_S.gguf) | Q4_K_S | 40.347 GB | small, greater quality loss |
156
+ | [Llama-3-70B-Orpo-v0.1-Q4_K_M.gguf](https://huggingface.co/tensorblock/Llama-3-70B-Orpo-v0.1-GGUF/blob/main/Llama-3-70B-Orpo-v0.1-Q4_K_M.gguf) | Q4_K_M | 42.520 GB | medium, balanced quality - recommended |
157
+ | [Llama-3-70B-Orpo-v0.1-Q5_0.gguf](https://huggingface.co/tensorblock/Llama-3-70B-Orpo-v0.1-GGUF/blob/main/Llama-3-70B-Orpo-v0.1-Q5_0.gguf) | Q5_0 | 48.657 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
158
+ | [Llama-3-70B-Orpo-v0.1-Q5_K_S.gguf](https://huggingface.co/tensorblock/Llama-3-70B-Orpo-v0.1-GGUF/blob/main/Llama-3-70B-Orpo-v0.1-Q5_K_S.gguf) | Q5_K_S | 48.657 GB | large, low quality loss - recommended |
159
+ | [Llama-3-70B-Orpo-v0.1-Q5_K_M.gguf](https://huggingface.co/tensorblock/Llama-3-70B-Orpo-v0.1-GGUF/blob/main/Llama-3-70B-Orpo-v0.1-Q5_K_M.gguf) | Q5_K_M | 49.950 GB | large, very low quality loss - recommended |
160
+ | [Llama-3-70B-Orpo-v0.1-Q6_K](https://huggingface.co/tensorblock/Llama-3-70B-Orpo-v0.1-GGUF/blob/main/Llama-3-70B-Orpo-v0.1-Q6_K) | Q6_K | 57.888 GB | very large, extremely low quality loss |
161
+ | [Llama-3-70B-Orpo-v0.1-Q8_0](https://huggingface.co/tensorblock/Llama-3-70B-Orpo-v0.1-GGUF/blob/main/Llama-3-70B-Orpo-v0.1-Q8_0) | Q8_0 | 74.975 GB | very large, extremely low quality loss - not recommended |
162
+
163
+
164
+ ## Downloading instruction
165
+
166
+ ### Command line
167
+
168
+ Firstly, install Huggingface Client
169
+
170
+ ```shell
171
+ pip install -U "huggingface_hub[cli]"
172
+ ```
173
+
174
+ Then, downoad the individual model file the a local directory
175
+
176
+ ```shell
177
+ huggingface-cli download tensorblock/Llama-3-70B-Orpo-v0.1-GGUF --include "Llama-3-70B-Orpo-v0.1-Q2_K.gguf" --local-dir MY_LOCAL_DIR
178
+ ```
179
+
180
+ If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf`), you can try:
181
+
182
+ ```shell
183
+ huggingface-cli download tensorblock/Llama-3-70B-Orpo-v0.1-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
184
+ ```