blockblockblock
commited on
Commit
•
f69a064
1
Parent(s):
3f7c914
Upload folder using huggingface_hub
Browse files- README.md +2 -1
- output.safetensors +2 -2
- tokenizer.json +6 -6
- tokenizer_config.json +6 -6
README.md
CHANGED
@@ -12,6 +12,7 @@ tags:
|
|
12 |
- distillation
|
13 |
- function calling
|
14 |
- json mode
|
|
|
15 |
model-index:
|
16 |
- name: Hermes-2-Pro-Llama-3-8B
|
17 |
results: []
|
@@ -262,7 +263,7 @@ import bitsandbytes, flash_attn
|
|
262 |
|
263 |
tokenizer = AutoTokenizer.from_pretrained('NousResearch/Hermes-2-Pro-Llama-3-8B', trust_remote_code=True)
|
264 |
model = LlamaForCausalLM.from_pretrained(
|
265 |
-
"Hermes-2-Pro-Llama-3-8B",
|
266 |
torch_dtype=torch.float16,
|
267 |
device_map="auto",
|
268 |
load_in_8bit=False,
|
|
|
12 |
- distillation
|
13 |
- function calling
|
14 |
- json mode
|
15 |
+
- axolotl
|
16 |
model-index:
|
17 |
- name: Hermes-2-Pro-Llama-3-8B
|
18 |
results: []
|
|
|
263 |
|
264 |
tokenizer = AutoTokenizer.from_pretrained('NousResearch/Hermes-2-Pro-Llama-3-8B', trust_remote_code=True)
|
265 |
model = LlamaForCausalLM.from_pretrained(
|
266 |
+
"NousResearch/Hermes-2-Pro-Llama-3-8B",
|
267 |
torch_dtype=torch.float16,
|
268 |
device_map="auto",
|
269 |
load_in_8bit=False,
|
output.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:610fc8421e0e9fa6c13e726f047f302799b6d67e5faa17cf435979e8ba35bce8
|
3 |
+
size 3431154416
|
tokenizer.json
CHANGED
@@ -46,7 +46,7 @@
|
|
46 |
"lstrip": false,
|
47 |
"rstrip": false,
|
48 |
"normalized": false,
|
49 |
-
"special":
|
50 |
},
|
51 |
{
|
52 |
"id": 128005,
|
@@ -55,7 +55,7 @@
|
|
55 |
"lstrip": false,
|
56 |
"rstrip": false,
|
57 |
"normalized": false,
|
58 |
-
"special":
|
59 |
},
|
60 |
{
|
61 |
"id": 128006,
|
@@ -82,7 +82,7 @@
|
|
82 |
"lstrip": false,
|
83 |
"rstrip": false,
|
84 |
"normalized": false,
|
85 |
-
"special":
|
86 |
},
|
87 |
{
|
88 |
"id": 128009,
|
@@ -100,7 +100,7 @@
|
|
100 |
"lstrip": false,
|
101 |
"rstrip": false,
|
102 |
"normalized": false,
|
103 |
-
"special":
|
104 |
},
|
105 |
{
|
106 |
"id": 128011,
|
@@ -109,7 +109,7 @@
|
|
109 |
"lstrip": false,
|
110 |
"rstrip": false,
|
111 |
"normalized": false,
|
112 |
-
"special":
|
113 |
},
|
114 |
{
|
115 |
"id": 128012,
|
@@ -118,7 +118,7 @@
|
|
118 |
"lstrip": false,
|
119 |
"rstrip": false,
|
120 |
"normalized": false,
|
121 |
-
"special":
|
122 |
},
|
123 |
{
|
124 |
"id": 128013,
|
|
|
46 |
"lstrip": false,
|
47 |
"rstrip": false,
|
48 |
"normalized": false,
|
49 |
+
"special": false
|
50 |
},
|
51 |
{
|
52 |
"id": 128005,
|
|
|
55 |
"lstrip": false,
|
56 |
"rstrip": false,
|
57 |
"normalized": false,
|
58 |
+
"special": false
|
59 |
},
|
60 |
{
|
61 |
"id": 128006,
|
|
|
82 |
"lstrip": false,
|
83 |
"rstrip": false,
|
84 |
"normalized": false,
|
85 |
+
"special": false
|
86 |
},
|
87 |
{
|
88 |
"id": 128009,
|
|
|
100 |
"lstrip": false,
|
101 |
"rstrip": false,
|
102 |
"normalized": false,
|
103 |
+
"special": false
|
104 |
},
|
105 |
{
|
106 |
"id": 128011,
|
|
|
109 |
"lstrip": false,
|
110 |
"rstrip": false,
|
111 |
"normalized": false,
|
112 |
+
"special": false
|
113 |
},
|
114 |
{
|
115 |
"id": 128012,
|
|
|
118 |
"lstrip": false,
|
119 |
"rstrip": false,
|
120 |
"normalized": false,
|
121 |
+
"special": false
|
122 |
},
|
123 |
{
|
124 |
"id": 128013,
|
tokenizer_config.json
CHANGED
@@ -38,7 +38,7 @@
|
|
38 |
"normalized": false,
|
39 |
"rstrip": false,
|
40 |
"single_word": false,
|
41 |
-
"special":
|
42 |
},
|
43 |
"128005": {
|
44 |
"content": "<tool_response>",
|
@@ -46,7 +46,7 @@
|
|
46 |
"normalized": false,
|
47 |
"rstrip": false,
|
48 |
"single_word": false,
|
49 |
-
"special":
|
50 |
},
|
51 |
"128006": {
|
52 |
"content": "<|start_header_id|>",
|
@@ -70,7 +70,7 @@
|
|
70 |
"normalized": false,
|
71 |
"rstrip": false,
|
72 |
"single_word": false,
|
73 |
-
"special":
|
74 |
},
|
75 |
"128009": {
|
76 |
"content": "<|eot_id>",
|
@@ -86,7 +86,7 @@
|
|
86 |
"normalized": false,
|
87 |
"rstrip": false,
|
88 |
"single_word": false,
|
89 |
-
"special":
|
90 |
},
|
91 |
"128011": {
|
92 |
"content": "</tool_call>",
|
@@ -94,7 +94,7 @@
|
|
94 |
"normalized": false,
|
95 |
"rstrip": false,
|
96 |
"single_word": false,
|
97 |
-
"special":
|
98 |
},
|
99 |
"128012": {
|
100 |
"content": "</tool_response>",
|
@@ -102,7 +102,7 @@
|
|
102 |
"normalized": false,
|
103 |
"rstrip": false,
|
104 |
"single_word": false,
|
105 |
-
"special":
|
106 |
},
|
107 |
"128013": {
|
108 |
"content": "<|reserved_special_token_8|>",
|
|
|
38 |
"normalized": false,
|
39 |
"rstrip": false,
|
40 |
"single_word": false,
|
41 |
+
"special": false
|
42 |
},
|
43 |
"128005": {
|
44 |
"content": "<tool_response>",
|
|
|
46 |
"normalized": false,
|
47 |
"rstrip": false,
|
48 |
"single_word": false,
|
49 |
+
"special": false
|
50 |
},
|
51 |
"128006": {
|
52 |
"content": "<|start_header_id|>",
|
|
|
70 |
"normalized": false,
|
71 |
"rstrip": false,
|
72 |
"single_word": false,
|
73 |
+
"special": false
|
74 |
},
|
75 |
"128009": {
|
76 |
"content": "<|eot_id>",
|
|
|
86 |
"normalized": false,
|
87 |
"rstrip": false,
|
88 |
"single_word": false,
|
89 |
+
"special": false
|
90 |
},
|
91 |
"128011": {
|
92 |
"content": "</tool_call>",
|
|
|
94 |
"normalized": false,
|
95 |
"rstrip": false,
|
96 |
"single_word": false,
|
97 |
+
"special": false
|
98 |
},
|
99 |
"128012": {
|
100 |
"content": "</tool_response>",
|
|
|
102 |
"normalized": false,
|
103 |
"rstrip": false,
|
104 |
"single_word": false,
|
105 |
+
"special": false
|
106 |
},
|
107 |
"128013": {
|
108 |
"content": "<|reserved_special_token_8|>",
|