cappuch commited on
Commit
5f4a958
1 Parent(s): 68ed8e4

Upload 5 files

Browse files
config.json CHANGED
@@ -1,18 +1,17 @@
1
  {
2
- "_name_or_path": "unsloth/llama-3-8b-bnb-4bit",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "bos_token_id": 128000,
9
- "eos_token_id": 128001,
10
  "hidden_act": "silu",
11
  "hidden_size": 4096,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 14336,
14
  "max_position_embeddings": 8192,
15
- "mlp_bias": false,
16
  "model_type": "llama",
17
  "num_attention_heads": 32,
18
  "num_hidden_layers": 32,
@@ -22,9 +21,8 @@
22
  "rope_scaling": null,
23
  "rope_theta": 500000.0,
24
  "tie_word_embeddings": false,
25
- "torch_dtype": "float16",
26
- "transformers_version": "4.41.0",
27
- "unsloth_version": "2024.5",
28
  "use_cache": true,
29
  "vocab_size": 128256
30
  }
 
1
  {
2
+ "_name_or_path": "goddardexperiments/HermesProInstructV10",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "bos_token_id": 128000,
9
+ "eos_token_id": 128003,
10
  "hidden_act": "silu",
11
  "hidden_size": 4096,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 14336,
14
  "max_position_embeddings": 8192,
 
15
  "model_type": "llama",
16
  "num_attention_heads": 32,
17
  "num_hidden_layers": 32,
 
21
  "rope_scaling": null,
22
  "rope_theta": 500000.0,
23
  "tie_word_embeddings": false,
24
+ "torch_dtype": "bfloat16",
25
+ "transformers_version": "4.40.0.dev0",
 
26
  "use_cache": true,
27
  "vocab_size": 128256
28
  }
generation_config.json CHANGED
@@ -1,9 +1,8 @@
1
  {
 
2
  "bos_token_id": 128000,
3
  "do_sample": true,
4
- "eos_token_id": 128001,
5
- "max_length": 4096,
6
- "temperature": 0.6,
7
- "top_p": 0.9,
8
- "transformers_version": "4.41.0"
9
  }
 
1
  {
2
+ "_from_model_config": true,
3
  "bos_token_id": 128000,
4
  "do_sample": true,
5
+ "eos_token_id": 128003,
6
+ "transformers_version": "4.40.0.dev0",
7
+ "use_cache": false
 
 
8
  }
special_tokens_map.json CHANGED
@@ -7,14 +7,14 @@
7
  "single_word": false
8
  },
9
  "eos_token": {
10
- "content": "<|end_of_text|>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
  "pad_token": {
17
- "content": "<|reserved_special_token_250|>",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
 
7
  "single_word": false
8
  },
9
  "eos_token": {
10
+ "content": "<|im_end|>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
  "pad_token": {
17
+ "content": "<|end_of_text|>",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
tokenizer.json CHANGED
@@ -23,7 +23,7 @@
23
  },
24
  {
25
  "id": 128002,
26
- "content": "<|reserved_special_token_0|>",
27
  "single_word": false,
28
  "lstrip": false,
29
  "rstrip": false,
@@ -32,7 +32,7 @@
32
  },
33
  {
34
  "id": 128003,
35
- "content": "<|reserved_special_token_1|>",
36
  "single_word": false,
37
  "lstrip": false,
38
  "rstrip": false,
@@ -41,21 +41,21 @@
41
  },
42
  {
43
  "id": 128004,
44
- "content": "<|reserved_special_token_2|>",
45
  "single_word": false,
46
  "lstrip": false,
47
  "rstrip": false,
48
  "normalized": false,
49
- "special": true
50
  },
51
  {
52
  "id": 128005,
53
- "content": "<|reserved_special_token_3|>",
54
  "single_word": false,
55
  "lstrip": false,
56
  "rstrip": false,
57
  "normalized": false,
58
- "special": true
59
  },
60
  {
61
  "id": 128006,
@@ -77,12 +77,12 @@
77
  },
78
  {
79
  "id": 128008,
80
- "content": "<|reserved_special_token_4|>",
81
  "single_word": false,
82
  "lstrip": false,
83
  "rstrip": false,
84
  "normalized": false,
85
- "special": true
86
  },
87
  {
88
  "id": 128009,
@@ -95,30 +95,30 @@
95
  },
96
  {
97
  "id": 128010,
98
- "content": "<|reserved_special_token_5|>",
99
  "single_word": false,
100
  "lstrip": false,
101
  "rstrip": false,
102
  "normalized": false,
103
- "special": true
104
  },
105
  {
106
  "id": 128011,
107
- "content": "<|reserved_special_token_6|>",
108
  "single_word": false,
109
  "lstrip": false,
110
  "rstrip": false,
111
  "normalized": false,
112
- "special": true
113
  },
114
  {
115
  "id": 128012,
116
- "content": "<|reserved_special_token_7|>",
117
  "single_word": false,
118
  "lstrip": false,
119
  "rstrip": false,
120
  "normalized": false,
121
- "special": true
122
  },
123
  {
124
  "id": 128013,
@@ -2407,7 +2407,6 @@
2407
  "end_of_word_suffix": null,
2408
  "fuse_unk": false,
2409
  "byte_fallback": false,
2410
- "ignore_merges": true,
2411
  "vocab": {
2412
  "!": 0,
2413
  "\"": 1,
 
23
  },
24
  {
25
  "id": 128002,
26
+ "content": "<|im_start|>",
27
  "single_word": false,
28
  "lstrip": false,
29
  "rstrip": false,
 
32
  },
33
  {
34
  "id": 128003,
35
+ "content": "<|im_end|>",
36
  "single_word": false,
37
  "lstrip": false,
38
  "rstrip": false,
 
41
  },
42
  {
43
  "id": 128004,
44
+ "content": "<tool_call>",
45
  "single_word": false,
46
  "lstrip": false,
47
  "rstrip": false,
48
  "normalized": false,
49
+ "special": false
50
  },
51
  {
52
  "id": 128005,
53
+ "content": "<tool_response>",
54
  "single_word": false,
55
  "lstrip": false,
56
  "rstrip": false,
57
  "normalized": false,
58
+ "special": false
59
  },
60
  {
61
  "id": 128006,
 
77
  },
78
  {
79
  "id": 128008,
80
+ "content": "<tools>",
81
  "single_word": false,
82
  "lstrip": false,
83
  "rstrip": false,
84
  "normalized": false,
85
+ "special": false
86
  },
87
  {
88
  "id": 128009,
 
95
  },
96
  {
97
  "id": 128010,
98
+ "content": "</tools>",
99
  "single_word": false,
100
  "lstrip": false,
101
  "rstrip": false,
102
  "normalized": false,
103
+ "special": false
104
  },
105
  {
106
  "id": 128011,
107
+ "content": "</tool_call>",
108
  "single_word": false,
109
  "lstrip": false,
110
  "rstrip": false,
111
  "normalized": false,
112
+ "special": false
113
  },
114
  {
115
  "id": 128012,
116
+ "content": "</tool_response>",
117
  "single_word": false,
118
  "lstrip": false,
119
  "rstrip": false,
120
  "normalized": false,
121
+ "special": false
122
  },
123
  {
124
  "id": 128013,
 
2407
  "end_of_word_suffix": null,
2408
  "fuse_unk": false,
2409
  "byte_fallback": false,
 
2410
  "vocab": {
2411
  "!": 0,
2412
  "\"": 1,
tokenizer_config.json CHANGED
@@ -17,7 +17,7 @@
17
  "special": true
18
  },
19
  "128002": {
20
- "content": "<|reserved_special_token_0|>",
21
  "lstrip": false,
22
  "normalized": false,
23
  "rstrip": false,
@@ -25,7 +25,7 @@
25
  "special": true
26
  },
27
  "128003": {
28
- "content": "<|reserved_special_token_1|>",
29
  "lstrip": false,
30
  "normalized": false,
31
  "rstrip": false,
@@ -33,20 +33,20 @@
33
  "special": true
34
  },
35
  "128004": {
36
- "content": "<|reserved_special_token_2|>",
37
  "lstrip": false,
38
  "normalized": false,
39
  "rstrip": false,
40
  "single_word": false,
41
- "special": true
42
  },
43
  "128005": {
44
- "content": "<|reserved_special_token_3|>",
45
  "lstrip": false,
46
  "normalized": false,
47
  "rstrip": false,
48
  "single_word": false,
49
- "special": true
50
  },
51
  "128006": {
52
  "content": "<|start_header_id|>",
@@ -65,12 +65,12 @@
65
  "special": true
66
  },
67
  "128008": {
68
- "content": "<|reserved_special_token_4|>",
69
  "lstrip": false,
70
  "normalized": false,
71
  "rstrip": false,
72
  "single_word": false,
73
- "special": true
74
  },
75
  "128009": {
76
  "content": "<|eot_id|>",
@@ -81,28 +81,28 @@
81
  "special": true
82
  },
83
  "128010": {
84
- "content": "<|reserved_special_token_5|>",
85
  "lstrip": false,
86
  "normalized": false,
87
  "rstrip": false,
88
  "single_word": false,
89
- "special": true
90
  },
91
  "128011": {
92
- "content": "<|reserved_special_token_6|>",
93
  "lstrip": false,
94
  "normalized": false,
95
  "rstrip": false,
96
  "single_word": false,
97
- "special": true
98
  },
99
  "128012": {
100
- "content": "<|reserved_special_token_7|>",
101
  "lstrip": false,
102
  "normalized": false,
103
  "rstrip": false,
104
  "single_word": false,
105
- "special": true
106
  },
107
  "128013": {
108
  "content": "<|reserved_special_token_8|>",
@@ -2050,14 +2050,14 @@
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
 
2053
  "clean_up_tokenization_spaces": true,
2054
- "eos_token": "<|end_of_text|>",
2055
  "model_input_names": [
2056
  "input_ids",
2057
  "attention_mask"
2058
  ],
2059
- "model_max_length": 8192,
2060
- "pad_token": "<|reserved_special_token_250|>",
2061
- "padding_side": "left",
2062
  "tokenizer_class": "PreTrainedTokenizerFast"
2063
  }
 
17
  "special": true
18
  },
19
  "128002": {
20
+ "content": "<|im_start|>",
21
  "lstrip": false,
22
  "normalized": false,
23
  "rstrip": false,
 
25
  "special": true
26
  },
27
  "128003": {
28
+ "content": "<|im_end|>",
29
  "lstrip": false,
30
  "normalized": false,
31
  "rstrip": false,
 
33
  "special": true
34
  },
35
  "128004": {
36
+ "content": "<tool_call>",
37
  "lstrip": false,
38
  "normalized": false,
39
  "rstrip": false,
40
  "single_word": false,
41
+ "special": false
42
  },
43
  "128005": {
44
+ "content": "<tool_response>",
45
  "lstrip": false,
46
  "normalized": false,
47
  "rstrip": false,
48
  "single_word": false,
49
+ "special": false
50
  },
51
  "128006": {
52
  "content": "<|start_header_id|>",
 
65
  "special": true
66
  },
67
  "128008": {
68
+ "content": "<tools>",
69
  "lstrip": false,
70
  "normalized": false,
71
  "rstrip": false,
72
  "single_word": false,
73
+ "special": false
74
  },
75
  "128009": {
76
  "content": "<|eot_id|>",
 
81
  "special": true
82
  },
83
  "128010": {
84
+ "content": "</tools>",
85
  "lstrip": false,
86
  "normalized": false,
87
  "rstrip": false,
88
  "single_word": false,
89
+ "special": false
90
  },
91
  "128011": {
92
+ "content": "</tool_call>",
93
  "lstrip": false,
94
  "normalized": false,
95
  "rstrip": false,
96
  "single_word": false,
97
+ "special": false
98
  },
99
  "128012": {
100
+ "content": "</tool_response>",
101
  "lstrip": false,
102
  "normalized": false,
103
  "rstrip": false,
104
  "single_word": false,
105
+ "special": false
106
  },
107
  "128013": {
108
  "content": "<|reserved_special_token_8|>",
 
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
+ "chat_template": "{{bos_token}}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
2054
  "clean_up_tokenization_spaces": true,
2055
+ "eos_token": "<|im_end|>",
2056
  "model_input_names": [
2057
  "input_ids",
2058
  "attention_mask"
2059
  ],
2060
+ "model_max_length": 1000000000000000019884624838656,
2061
+ "pad_token": "<|end_of_text|>",
 
2062
  "tokenizer_class": "PreTrainedTokenizerFast"
2063
  }