chrisociepa commited on
Commit
7a52e4b
1 Parent(s): 2a5de68

Fix configs

Browse files
added_tokens.json ADDED
@@ -0,0 +1,130 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<|control_100|>": 32099,
3
+ "<|control_101|>": 32100,
4
+ "<|control_102|>": 32101,
5
+ "<|control_103|>": 32102,
6
+ "<|control_104|>": 32103,
7
+ "<|control_105|>": 32104,
8
+ "<|control_106|>": 32105,
9
+ "<|control_107|>": 32106,
10
+ "<|control_108|>": 32107,
11
+ "<|control_109|>": 32108,
12
+ "<|control_10|>": 32009,
13
+ "<|control_110|>": 32109,
14
+ "<|control_111|>": 32110,
15
+ "<|control_112|>": 32111,
16
+ "<|control_113|>": 32112,
17
+ "<|control_114|>": 32113,
18
+ "<|control_115|>": 32114,
19
+ "<|control_116|>": 32115,
20
+ "<|control_117|>": 32116,
21
+ "<|control_118|>": 32117,
22
+ "<|control_119|>": 32118,
23
+ "<|control_11|>": 32010,
24
+ "<|control_120|>": 32119,
25
+ "<|control_121|>": 32120,
26
+ "<|control_122|>": 32121,
27
+ "<|control_123|>": 32122,
28
+ "<|control_124|>": 32123,
29
+ "<|control_125|>": 32124,
30
+ "<|control_126|>": 32125,
31
+ "<|control_127|>": 32126,
32
+ "<|control_128|>": 32127,
33
+ "<|control_12|>": 32011,
34
+ "<|control_13|>": 32012,
35
+ "<|control_14|>": 32013,
36
+ "<|control_15|>": 32014,
37
+ "<|control_16|>": 32015,
38
+ "<|control_17|>": 32016,
39
+ "<|control_18|>": 32017,
40
+ "<|control_19|>": 32018,
41
+ "<|control_20|>": 32019,
42
+ "<|control_21|>": 32020,
43
+ "<|control_22|>": 32021,
44
+ "<|control_23|>": 32022,
45
+ "<|control_24|>": 32023,
46
+ "<|control_25|>": 32024,
47
+ "<|control_26|>": 32025,
48
+ "<|control_27|>": 32026,
49
+ "<|control_28|>": 32027,
50
+ "<|control_29|>": 32028,
51
+ "<|control_30|>": 32029,
52
+ "<|control_31|>": 32030,
53
+ "<|control_32|>": 32031,
54
+ "<|control_33|>": 32032,
55
+ "<|control_34|>": 32033,
56
+ "<|control_35|>": 32034,
57
+ "<|control_36|>": 32035,
58
+ "<|control_37|>": 32036,
59
+ "<|control_38|>": 32037,
60
+ "<|control_39|>": 32038,
61
+ "<|control_40|>": 32039,
62
+ "<|control_41|>": 32040,
63
+ "<|control_42|>": 32041,
64
+ "<|control_43|>": 32042,
65
+ "<|control_44|>": 32043,
66
+ "<|control_45|>": 32044,
67
+ "<|control_46|>": 32045,
68
+ "<|control_47|>": 32046,
69
+ "<|control_48|>": 32047,
70
+ "<|control_49|>": 32048,
71
+ "<|control_50|>": 32049,
72
+ "<|control_51|>": 32050,
73
+ "<|control_52|>": 32051,
74
+ "<|control_53|>": 32052,
75
+ "<|control_54|>": 32053,
76
+ "<|control_55|>": 32054,
77
+ "<|control_56|>": 32055,
78
+ "<|control_57|>": 32056,
79
+ "<|control_58|>": 32057,
80
+ "<|control_59|>": 32058,
81
+ "<|control_60|>": 32059,
82
+ "<|control_61|>": 32060,
83
+ "<|control_62|>": 32061,
84
+ "<|control_63|>": 32062,
85
+ "<|control_64|>": 32063,
86
+ "<|control_65|>": 32064,
87
+ "<|control_66|>": 32065,
88
+ "<|control_67|>": 32066,
89
+ "<|control_68|>": 32067,
90
+ "<|control_69|>": 32068,
91
+ "<|control_6|>": 32005,
92
+ "<|control_70|>": 32069,
93
+ "<|control_71|>": 32070,
94
+ "<|control_72|>": 32071,
95
+ "<|control_73|>": 32072,
96
+ "<|control_74|>": 32073,
97
+ "<|control_75|>": 32074,
98
+ "<|control_76|>": 32075,
99
+ "<|control_77|>": 32076,
100
+ "<|control_78|>": 32077,
101
+ "<|control_79|>": 32078,
102
+ "<|control_7|>": 32006,
103
+ "<|control_80|>": 32079,
104
+ "<|control_81|>": 32080,
105
+ "<|control_82|>": 32081,
106
+ "<|control_83|>": 32082,
107
+ "<|control_84|>": 32083,
108
+ "<|control_85|>": 32084,
109
+ "<|control_86|>": 32085,
110
+ "<|control_87|>": 32086,
111
+ "<|control_88|>": 32087,
112
+ "<|control_89|>": 32088,
113
+ "<|control_8|>": 32007,
114
+ "<|control_90|>": 32089,
115
+ "<|control_91|>": 32090,
116
+ "<|control_92|>": 32091,
117
+ "<|control_93|>": 32092,
118
+ "<|control_94|>": 32093,
119
+ "<|control_95|>": 32094,
120
+ "<|control_96|>": 32095,
121
+ "<|control_97|>": 32096,
122
+ "<|control_98|>": 32097,
123
+ "<|control_99|>": 32098,
124
+ "<|control_9|>": 32008,
125
+ "<|function_call|>": 32004,
126
+ "<|function_list|>": 32002,
127
+ "<|function_output|>": 32003,
128
+ "<|im_end|>": 32001,
129
+ "<|im_start|>": 32000
130
+ }
config.json CHANGED
@@ -1,17 +1,15 @@
1
  {
2
- "_name_or_path": "/mnt/sda/llm/benchmarks/FastChat/fastchat/llm_judge/bielik_2_final/2.0/",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 1,
8
- "eos_token_id": 2,
9
- "head_dim": 128,
10
  "hidden_act": "silu",
11
  "hidden_size": 4096,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 14336,
14
- "max_position_embeddings": 8192,
15
  "model_type": "mistral",
16
  "num_attention_heads": 32,
17
  "num_hidden_layers": 50,
 
1
  {
 
2
  "architectures": [
3
  "MistralForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
  "bos_token_id": 1,
7
+ "eos_token_id": 32001,
 
8
  "hidden_act": "silu",
9
  "hidden_size": 4096,
10
  "initializer_range": 0.02,
11
  "intermediate_size": 14336,
12
+ "max_position_embeddings": 32768,
13
  "model_type": "mistral",
14
  "num_attention_heads": 32,
15
  "num_hidden_layers": 50,
generation_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 32001,
5
+ "pad_token": 2,
6
+ "unk_token": 0,
7
+ "transformers_version": "4.39.3"
8
+ }
special_tokens_map.json CHANGED
@@ -137,7 +137,7 @@
137
  "single_word": false
138
  },
139
  "eos_token": {
140
- "content": "</s>",
141
  "lstrip": false,
142
  "normalized": false,
143
  "rstrip": false,
 
137
  "single_word": false
138
  },
139
  "eos_token": {
140
+ "content": "<|im_end|>",
141
  "lstrip": false,
142
  "normalized": false,
143
  "rstrip": false,
tokenizer_config.json CHANGED
@@ -1185,7 +1185,7 @@
1185
  "bos_token": "<s>",
1186
  "chat_template": "{{bos_token}}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
1187
  "clean_up_tokenization_spaces": true,
1188
- "eos_token": "</s>",
1189
  "legacy": true,
1190
  "model_max_length": 1000000000000000019884624838656,
1191
  "pad_token": "</s>",
 
1185
  "bos_token": "<s>",
1186
  "chat_template": "{{bos_token}}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
1187
  "clean_up_tokenization_spaces": true,
1188
+ "eos_token": "<|im_end|>",
1189
  "legacy": true,
1190
  "model_max_length": 1000000000000000019884624838656,
1191
  "pad_token": "</s>",