bobtk commited on
Commit
7f8c3e1
1 Parent(s): 36a61d5

Upload folder using huggingface_hub (#2)

Browse files

- 494511499df1b847e1ad5cdf4b7277abbeaca1780e811eec16ab0df38fa6f459 (5a789cb24ce58790490ecb492141d2b67a97ac27)
- f5385956b6d841b83eadbad3008c7208b50dd630b049ee17dda319669f94d970 (ac6bbbc29ce967d9548b2e1c60d59b74f87d5a9c)
- a3e25a8612e9cc6d80f5308645ccad6930b4a27ed4217bcce8f8cea1d8c66275 (cc354ebabef5ca44431d0c1b143dc84e63b5f683)
- 092c4d5dec998d6efb55f3b80537c849b422d489cda7d759e6f704949331a3e8 (c890d5e17349152394199957f1196d0549b4eab6)
- ed6b952624812329c74b07b674d57d72f2b12cfaed23577484a7f126452ac6f5 (1153aa78397f778ba728338dbf3743a72fdde2ad)
- 43b09596998877dadb2ec3d28b3e97cf61707c4291d9c3cf68645aaa4ba54f14 (42086adf71c9a8b820be2c6cc8445e9445b17ab7)
- 3bb44929deab8253c334c1c05d95e28bbf177ff5ff37d459024f54e08e2311fb (cf96b819290e7b0c817bb4375e0673bc993b35d1)
- db2e4f3671276786afa8f4c64b639acfc55a7f23fbfd58e549e299091265d7a5 (dc99bdfe6989ffb55a9a376dce0acd7cfbf10ab6)
- 475ef5814b03d9361d0fe4402487544a1f62d3185b97792a3bad7d3bd5da42f3 (2565f15788e44e5cbbb905b8b15883e3814feb3d)
- 76e07337d43da4cd3b200179457f7c1c5ff63b8c639b283e4ff9b9ed5fcec5ab (c2fb3ef3c0885b8fbedc0f28b3c22b0ae58e5b7f)
- 69474179abe83b069ff54f5c40540a4a2665ca68c0809ab23311e07c763bff66 (7b0f7476117c96d15908e59762791defe997b95b)
- a2a8be890965d8af7b45b63f133dea0c70951943d84a1e1633dbf1120dde634b (098a7def9269e9879ba9ba1b0e1f5611f977a27d)
- 4eaa08f7ef01794229363b8b3ceea46b12cc7d5ae24252c28f7c310cb8ccce58 (3704860c017690f214cd843b0625f6199da8a2eb)
- cbb20ccda3a46bae86e8b698a4b8a01d82f8719a89b98e70fb40ddeb14b052b1 (fac66296d45d18bac26f21d489b6e00697267276)
- 5a31daae8e60c36371a7fc39e708d4d5475e3434974095f0bf01481232d65888 (aa627c83cee0010f2871de402cab4508d296274c)
- fea5711ea55ed95ef2435c9c6d8b35bf5d377e6de05056ab2e90df53314c05ed (be92b67ddc7de80c8e8dece88d898ccd35c92cc2)
- 06e9b59389dcb5c7964660c260bced6f6e2bd2b7470f620f56d470ef30e4056a (60cb48b014b81f1c8f95e1efbd07860e8ed1c773)
- 1d2049f0879097ecac7298dd4e2a18738278b37847afadd7b200da52f69220f5 (c243f3b1a3f0001392fdecd7f489c579202273c7)
- ca199ff28b7a4f041d904321f813e714071b5cd4bdcd2971b73b4aabcce11ef9 (d0d80c6c1f4d05152547259a3f26c91ec6c5cd45)
- 8bba90f280c9064ff3d9132c65387df182489bd854fc256562c232dd96335221 (1565ce25aed0e064ef52348c2aa0aa0db1d00e39)
- feb36d683e63639e9aff19bb045a5f22a1cde04fd056c6e557f00e7468dd9c97 (363e2b439faa57b6c49a8c5925dcd762ee145cf1)
- 2429157291d10131439d7e287f3e664c50ba7d17f7cfd465f33a3c2e89be43b7 (c39cbffb728e0468a77a4f1163e2c2adacec6b7b)
- 711c27ea8b420520a6fd185d4fce0dfb84f3a02ce6f668e1f0a733f8a8123100 (8bad67379fb4578f0867514a22bd411d858c476b)

.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,292 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - fr
5
+ - de
6
+ - es
7
+ - it
8
+ - pt
9
+ - ja
10
+ - ko
11
+ - zh
12
+ - ar
13
+ library_name: transformers
14
+ license: cc-by-nc-4.0
15
+ tags:
16
+ - mlx
17
+ inference: false
18
+ extra_gated_prompt: By submitting this form, you agree to the [License Agreement](https://cohere.com/c4ai-cc-by-nc-license) and
19
+ acknowledge that the information you provide will be collected, used, and shared
20
+ in accordance with Cohere’s [Privacy Policy]( https://cohere.com/privacy).
21
+ extra_gated_fields:
22
+ Name: text
23
+ Affiliation: text
24
+ Country:
25
+ type: select
26
+ options:
27
+ - Aruba
28
+ - Afghanistan
29
+ - Angola
30
+ - Anguilla
31
+ - Åland Islands
32
+ - Albania
33
+ - Andorra
34
+ - United Arab Emirates
35
+ - Argentina
36
+ - Armenia
37
+ - American Samoa
38
+ - Antarctica
39
+ - French Southern Territories
40
+ - Antigua and Barbuda
41
+ - Australia
42
+ - Austria
43
+ - Azerbaijan
44
+ - Burundi
45
+ - Belgium
46
+ - Benin
47
+ - Bonaire Sint Eustatius and Saba
48
+ - Burkina Faso
49
+ - Bangladesh
50
+ - Bulgaria
51
+ - Bahrain
52
+ - Bahamas
53
+ - Bosnia and Herzegovina
54
+ - Saint Barthélemy
55
+ - Belarus
56
+ - Belize
57
+ - Bermuda
58
+ - Plurinational State of Bolivia
59
+ - Brazil
60
+ - Barbados
61
+ - Brunei-Darussalam
62
+ - Bhutan
63
+ - Bouvet-Island
64
+ - Botswana
65
+ - Central African Republic
66
+ - Canada
67
+ - Cocos (Keeling) Islands
68
+ - Switzerland
69
+ - Chile
70
+ - China
71
+ - Côte-dIvoire
72
+ - Cameroon
73
+ - Democratic Republic of the Congo
74
+ - Cook Islands
75
+ - Colombia
76
+ - Comoros
77
+ - Cabo Verde
78
+ - Costa Rica
79
+ - Cuba
80
+ - Curaçao
81
+ - Christmas Island
82
+ - Cayman Islands
83
+ - Cyprus
84
+ - Czechia
85
+ - Germany
86
+ - Djibouti
87
+ - Dominica
88
+ - Denmark
89
+ - Dominican Republic
90
+ - Algeria
91
+ - Ecuador
92
+ - Egypt
93
+ - Eritrea
94
+ - Western Sahara
95
+ - Spain
96
+ - Estonia
97
+ - Ethiopia
98
+ - Finland
99
+ - Fiji
100
+ - Falkland Islands (Malvinas)
101
+ - France
102
+ - Faroe Islands
103
+ - Federated States of Micronesia
104
+ - Gabon
105
+ - United Kingdom
106
+ - Georgia
107
+ - Guernsey
108
+ - Ghana
109
+ - Gibraltar
110
+ - Guinea
111
+ - Guadeloupe
112
+ - Gambia
113
+ - Guinea Bissau
114
+ - Equatorial Guinea
115
+ - Greece
116
+ - Grenada
117
+ - Greenland
118
+ - Guatemala
119
+ - French Guiana
120
+ - Guam
121
+ - Guyana
122
+ - Hong Kong
123
+ - Heard Island and McDonald Islands
124
+ - Honduras
125
+ - Croatia
126
+ - Haiti
127
+ - Hungary
128
+ - Indonesia
129
+ - Isle of Man
130
+ - India
131
+ - British Indian Ocean Territory
132
+ - Ireland
133
+ - Islamic Republic of Iran
134
+ - Iraq
135
+ - Iceland
136
+ - Israel
137
+ - Italy
138
+ - Jamaica
139
+ - Jersey
140
+ - Jordan
141
+ - Japan
142
+ - Kazakhstan
143
+ - Kenya
144
+ - Kyrgyzstan
145
+ - Cambodia
146
+ - Kiribati
147
+ - Saint-Kitts-and-Nevis
148
+ - South Korea
149
+ - Kuwait
150
+ - Lao-Peoples-Democratic-Republic
151
+ - Lebanon
152
+ - Liberia
153
+ - Libya
154
+ - Saint-Lucia
155
+ - Liechtenstein
156
+ - Sri Lanka
157
+ - Lesotho
158
+ - Lithuania
159
+ - Luxembourg
160
+ - Latvia
161
+ - Macao
162
+ - Saint Martin (French-part)
163
+ - Morocco
164
+ - Monaco
165
+ - Republic of Moldova
166
+ - Madagascar
167
+ - Maldives
168
+ - Mexico
169
+ - Marshall Islands
170
+ - North Macedonia
171
+ - Mali
172
+ - Malta
173
+ - Myanmar
174
+ - Montenegro
175
+ - Mongolia
176
+ - Northern Mariana Islands
177
+ - Mozambique
178
+ - Mauritania
179
+ - Montserrat
180
+ - Martinique
181
+ - Mauritius
182
+ - Malawi
183
+ - Malaysia
184
+ - Mayotte
185
+ - Namibia
186
+ - New Caledonia
187
+ - Niger
188
+ - Norfolk Island
189
+ - Nigeria
190
+ - Nicaragua
191
+ - Niue
192
+ - Netherlands
193
+ - Norway
194
+ - Nepal
195
+ - Nauru
196
+ - New Zealand
197
+ - Oman
198
+ - Pakistan
199
+ - Panama
200
+ - Pitcairn
201
+ - Peru
202
+ - Philippines
203
+ - Palau
204
+ - Papua New Guinea
205
+ - Poland
206
+ - Puerto Rico
207
+ - North Korea
208
+ - Portugal
209
+ - Paraguay
210
+ - State of Palestine
211
+ - French Polynesia
212
+ - Qatar
213
+ - Réunion
214
+ - Romania
215
+ - Russia
216
+ - Rwanda
217
+ - Saudi Arabia
218
+ - Sudan
219
+ - Senegal
220
+ - Singapore
221
+ - South Georgia and the South Sandwich Islands
222
+ - Saint Helena Ascension and Tristan da Cunha
223
+ - Svalbard and Jan Mayen
224
+ - Solomon Islands
225
+ - Sierra Leone
226
+ - El Salvador
227
+ - San Marino
228
+ - Somalia
229
+ - Saint Pierre and Miquelon
230
+ - Serbia
231
+ - South Sudan
232
+ - Sao Tome and Principe
233
+ - Suriname
234
+ - Slovakia
235
+ - Slovenia
236
+ - Sweden
237
+ - Eswatini
238
+ - Sint Maarten (Dutch-part)
239
+ - Seychelles
240
+ - Syrian Arab Republic
241
+ - Turks and Caicos Islands
242
+ - Chad
243
+ - Togo
244
+ - Thailand
245
+ - Tajikistan
246
+ - Tokelau
247
+ - Turkmenistan
248
+ - Timor Leste
249
+ - Tonga
250
+ - Trinidad and Tobago
251
+ - Tunisia
252
+ - Turkey
253
+ - Tuvalu
254
+ - Taiwan
255
+ - United Republic of Tanzania
256
+ - Uganda
257
+ - Ukraine
258
+ - United States Minor Outlying Islands
259
+ - Uruguay
260
+ - United-States
261
+ - Uzbekistan
262
+ - Holy See (Vatican City State)
263
+ - Saint Vincent and the Grenadines
264
+ - Bolivarian Republic of Venezuela
265
+ - Virgin Islands British
266
+ - Virgin Islands U.S.
267
+ - VietNam
268
+ - Vanuatu
269
+ - Wallis and Futuna
270
+ - Samoa
271
+ - Yemen
272
+ - South Africa
273
+ - Zambia
274
+ - Zimbabwe
275
+ ---
276
+
277
+ # mlx-community/c4ai-command-r-plus-08-2024-8bit
278
+
279
+ The Model [mlx-community/c4ai-command-r-plus-08-2024-8bit](https://huggingface.co/mlx-community/c4ai-command-r-plus-08-2024-8bit) was converted to MLX format from [CohereForAI/c4ai-command-r-plus-08-2024](https://huggingface.co/CohereForAI/c4ai-command-r-plus-08-2024) using mlx-lm version **0.18.1**.
280
+
281
+ ## Use with mlx
282
+
283
+ ```bash
284
+ pip install mlx-lm
285
+ ```
286
+
287
+ ```python
288
+ from mlx_lm import load, generate
289
+
290
+ model, tokenizer = load("mlx-community/c4ai-command-r-plus-08-2024-8bit")
291
+ response = generate(model, tokenizer, prompt="hello", verbose=True)
292
+ ```
config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "CohereForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 5,
8
+ "eos_token_id": 255001,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 12288,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 33792,
13
+ "layer_norm_eps": 1e-05,
14
+ "logit_scale": 0.8333333333333334,
15
+ "max_position_embeddings": 131072,
16
+ "model_type": "cohere",
17
+ "num_attention_heads": 96,
18
+ "num_hidden_layers": 64,
19
+ "num_key_value_heads": 8,
20
+ "pad_token_id": 0,
21
+ "quantization": {
22
+ "group_size": 64,
23
+ "bits": 8
24
+ },
25
+ "rope_theta": 8000000,
26
+ "torch_dtype": "float16",
27
+ "transformers_version": "4.44.0",
28
+ "use_cache": true,
29
+ "use_qk_norm": true,
30
+ "vocab_size": 256000
31
+ }
model-00001-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a96686e3555052f60bf64b6d502cd55c7ec7017c62a53395ce053861736fcae3
3
+ size 5361189410
model-00002-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70246a4d2d0308efd4f4e40798f81b0d5e6001c5c85abfba9062dd5b0c33b64b
3
+ size 5013665826
model-00003-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dc00955517f8b71a138030bab3238239cc94fc19d33eaa879f47c4541dfdfd9
3
+ size 5013665796
model-00004-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3a794fd97dc8c1658b1b1d631f336aa860e2b6c1360f52637f223271305dd5d
3
+ size 5013665834
model-00005-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:155ac76705bee9df17858cd6ec08c87b2eafc06df61bf0a2420f60d7439a6369
3
+ size 5013665854
model-00006-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69838a5e666b7b3964d66e95f7b33213d76884602e4865f7d7b869d3ba8ea403
3
+ size 5013665886
model-00007-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fcf19dd7788e84658874ca4d3e60dce4b4db0977799a74cdfd9a5c4c8d43220
3
+ size 5013665890
model-00008-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c96d129b30dfc64660c4ff7d267da692214c4bba31249b2baad6fb5f061a43f
3
+ size 5013665894
model-00009-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ac60294a746beaef03947dc8dd80a210128dbfd2bf3b0d46abdd3933480812f
3
+ size 5013665870
model-00010-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b5c52cc78be93a4e3d67ecfebd41920b2de71fc0c85de52c83ff3eea0f6e74f
3
+ size 5013665872
model-00011-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4195adc116449dc7ee2a954ca701f20de47681a648c4990e7675e8191bcced5
3
+ size 5013665878
model-00012-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:517d611cc115092392dbe93d46ac15f8175640214f0b0a730e73ca4572e97a4b
3
+ size 5013665866
model-00013-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ac86e819ff1caf3a2c360068e36ba5a53463d04e6823e4ee3ae10bcb44bddc1
3
+ size 5013665864
model-00014-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8b8bccffdd02639f7f00434e00c4f6de72ea8455d4eff6baa6a62f3f91fb694
3
+ size 5013665878
model-00015-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5413f88e01cd7155c5f07e9b2e2e6f57a17f6b9a5abf30bddf278828d437e8d6
3
+ size 5013665874
model-00016-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da917453ca1e4f9f6cc836602aa5e87def7e6647a5206d6280723ca93e6b8860
3
+ size 5013665854
model-00017-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c633ca02d95e097a766c0cda8eb106d7563bc1c9900dcf12f69318f2bdee10d
3
+ size 5013665878
model-00018-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14ea6b4a6268ad9c1b2e9757290caef7c289640890f99ef582e63c75e8211752
3
+ size 5013665878
model-00019-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28e8c439e30db1c76797b96c75e301b986aa2320361f56a4c69b8b2c90ff73dd
3
+ size 5013665888
model-00020-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e23ec5213950236911f8a0e4422a57d00e1ae8ee66ccda8b6d6e74b4ac21333
3
+ size 5013665874
model-00021-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd081a5a26dc2569e4ab2cbb4067190086028026ba778f3202f3e8336e9c67b8
3
+ size 5013665850
model-00022-of-00022.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27c0c2bb1433f193bcde40e32402142ea6a69f4066567f4f377e1ae4b93fd9cd
3
+ size 4666059360
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<BOS_TOKEN>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|END_OF_TURN_TOKEN|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<PAD>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c69a7ea6c0927dfac8c349186ebcf0466a4723c21cbdb2e850cf559f0bee92b8
3
+ size 12777433
tokenizer_config.json ADDED
@@ -0,0 +1,330 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": false,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<PAD>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<UNK>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "<CLS>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "3": {
31
+ "content": "<SEP>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ },
38
+ "4": {
39
+ "content": "<MASK_TOKEN>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": true
45
+ },
46
+ "5": {
47
+ "content": "<BOS_TOKEN>",
48
+ "lstrip": false,
49
+ "normalized": false,
50
+ "rstrip": false,
51
+ "single_word": false,
52
+ "special": true
53
+ },
54
+ "6": {
55
+ "content": "<EOS_TOKEN>",
56
+ "lstrip": false,
57
+ "normalized": false,
58
+ "rstrip": false,
59
+ "single_word": false,
60
+ "special": true
61
+ },
62
+ "7": {
63
+ "content": "<EOP_TOKEN>",
64
+ "lstrip": false,
65
+ "normalized": false,
66
+ "rstrip": false,
67
+ "single_word": false,
68
+ "special": true
69
+ },
70
+ "255000": {
71
+ "content": "<|START_OF_TURN_TOKEN|>",
72
+ "lstrip": false,
73
+ "normalized": false,
74
+ "rstrip": false,
75
+ "single_word": false,
76
+ "special": false
77
+ },
78
+ "255001": {
79
+ "content": "<|END_OF_TURN_TOKEN|>",
80
+ "lstrip": false,
81
+ "normalized": false,
82
+ "rstrip": false,
83
+ "single_word": false,
84
+ "special": true
85
+ },
86
+ "255002": {
87
+ "content": "<|YES_TOKEN|>",
88
+ "lstrip": false,
89
+ "normalized": false,
90
+ "rstrip": false,
91
+ "single_word": false,
92
+ "special": false
93
+ },
94
+ "255003": {
95
+ "content": "<|NO_TOKEN|>",
96
+ "lstrip": false,
97
+ "normalized": false,
98
+ "rstrip": false,
99
+ "single_word": false,
100
+ "special": false
101
+ },
102
+ "255004": {
103
+ "content": "<|GOOD_TOKEN|>",
104
+ "lstrip": false,
105
+ "normalized": false,
106
+ "rstrip": false,
107
+ "single_word": false,
108
+ "special": false
109
+ },
110
+ "255005": {
111
+ "content": "<|BAD_TOKEN|>",
112
+ "lstrip": false,
113
+ "normalized": false,
114
+ "rstrip": false,
115
+ "single_word": false,
116
+ "special": false
117
+ },
118
+ "255006": {
119
+ "content": "<|USER_TOKEN|>",
120
+ "lstrip": false,
121
+ "normalized": false,
122
+ "rstrip": false,
123
+ "single_word": false,
124
+ "special": false
125
+ },
126
+ "255007": {
127
+ "content": "<|CHATBOT_TOKEN|>",
128
+ "lstrip": false,
129
+ "normalized": false,
130
+ "rstrip": false,
131
+ "single_word": false,
132
+ "special": false
133
+ },
134
+ "255008": {
135
+ "content": "<|SYSTEM_TOKEN|>",
136
+ "lstrip": false,
137
+ "normalized": false,
138
+ "rstrip": false,
139
+ "single_word": false,
140
+ "special": false
141
+ },
142
+ "255009": {
143
+ "content": "<|USER_0_TOKEN|>",
144
+ "lstrip": false,
145
+ "normalized": false,
146
+ "rstrip": false,
147
+ "single_word": false,
148
+ "special": false
149
+ },
150
+ "255010": {
151
+ "content": "<|USER_1_TOKEN|>",
152
+ "lstrip": false,
153
+ "normalized": false,
154
+ "rstrip": false,
155
+ "single_word": false,
156
+ "special": false
157
+ },
158
+ "255011": {
159
+ "content": "<|USER_2_TOKEN|>",
160
+ "lstrip": false,
161
+ "normalized": false,
162
+ "rstrip": false,
163
+ "single_word": false,
164
+ "special": false
165
+ },
166
+ "255012": {
167
+ "content": "<|USER_3_TOKEN|>",
168
+ "lstrip": false,
169
+ "normalized": false,
170
+ "rstrip": false,
171
+ "single_word": false,
172
+ "special": false
173
+ },
174
+ "255013": {
175
+ "content": "<|USER_4_TOKEN|>",
176
+ "lstrip": false,
177
+ "normalized": false,
178
+ "rstrip": false,
179
+ "single_word": false,
180
+ "special": false
181
+ },
182
+ "255014": {
183
+ "content": "<|USER_5_TOKEN|>",
184
+ "lstrip": false,
185
+ "normalized": false,
186
+ "rstrip": false,
187
+ "single_word": false,
188
+ "special": false
189
+ },
190
+ "255015": {
191
+ "content": "<|USER_6_TOKEN|>",
192
+ "lstrip": false,
193
+ "normalized": false,
194
+ "rstrip": false,
195
+ "single_word": false,
196
+ "special": false
197
+ },
198
+ "255016": {
199
+ "content": "<|USER_7_TOKEN|>",
200
+ "lstrip": false,
201
+ "normalized": false,
202
+ "rstrip": false,
203
+ "single_word": false,
204
+ "special": false
205
+ },
206
+ "255017": {
207
+ "content": "<|USER_8_TOKEN|>",
208
+ "lstrip": false,
209
+ "normalized": false,
210
+ "rstrip": false,
211
+ "single_word": false,
212
+ "special": false
213
+ },
214
+ "255018": {
215
+ "content": "<|USER_9_TOKEN|>",
216
+ "lstrip": false,
217
+ "normalized": false,
218
+ "rstrip": false,
219
+ "single_word": false,
220
+ "special": false
221
+ },
222
+ "255019": {
223
+ "content": "<|EXTRA_0_TOKEN|>",
224
+ "lstrip": false,
225
+ "normalized": false,
226
+ "rstrip": false,
227
+ "single_word": false,
228
+ "special": false
229
+ },
230
+ "255020": {
231
+ "content": "<|EXTRA_1_TOKEN|>",
232
+ "lstrip": false,
233
+ "normalized": false,
234
+ "rstrip": false,
235
+ "single_word": false,
236
+ "special": false
237
+ },
238
+ "255021": {
239
+ "content": "<|EXTRA_2_TOKEN|>",
240
+ "lstrip": false,
241
+ "normalized": false,
242
+ "rstrip": false,
243
+ "single_word": false,
244
+ "special": false
245
+ },
246
+ "255022": {
247
+ "content": "<|EXTRA_3_TOKEN|>",
248
+ "lstrip": false,
249
+ "normalized": false,
250
+ "rstrip": false,
251
+ "single_word": false,
252
+ "special": false
253
+ },
254
+ "255023": {
255
+ "content": "<|EXTRA_4_TOKEN|>",
256
+ "lstrip": false,
257
+ "normalized": false,
258
+ "rstrip": false,
259
+ "single_word": false,
260
+ "special": false
261
+ },
262
+ "255024": {
263
+ "content": "<|EXTRA_5_TOKEN|>",
264
+ "lstrip": false,
265
+ "normalized": false,
266
+ "rstrip": false,
267
+ "single_word": false,
268
+ "special": false
269
+ },
270
+ "255025": {
271
+ "content": "<|EXTRA_6_TOKEN|>",
272
+ "lstrip": false,
273
+ "normalized": false,
274
+ "rstrip": false,
275
+ "single_word": false,
276
+ "special": false
277
+ },
278
+ "255026": {
279
+ "content": "<|EXTRA_7_TOKEN|>",
280
+ "lstrip": false,
281
+ "normalized": false,
282
+ "rstrip": false,
283
+ "single_word": false,
284
+ "special": false
285
+ },
286
+ "255027": {
287
+ "content": "<|EXTRA_8_TOKEN|>",
288
+ "lstrip": false,
289
+ "normalized": false,
290
+ "rstrip": false,
291
+ "single_word": false,
292
+ "special": false
293
+ },
294
+ "255028": {
295
+ "content": "<|EXTRA_9_TOKEN|>",
296
+ "lstrip": false,
297
+ "normalized": false,
298
+ "rstrip": false,
299
+ "single_word": false,
300
+ "special": false
301
+ }
302
+ },
303
+ "bos_token": "<BOS_TOKEN>",
304
+ "chat_template": [
305
+ {
306
+ "name": "default",
307
+ "template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% elif false == true %}{% set loop_messages = messages %}{% set system_message = 'You are a large language model called Command R+ built by the company Cohere. You act as a brilliant, sophisticated, AI-assistant chatbot trained to assist human users by providing thorough responses.' %}{% else %}{% set loop_messages = messages %}{% set system_message = false %}{% endif %}{% if system_message != false %}{{ '<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>' + system_message + '<|END_OF_TURN_TOKEN|>' }}{% endif %}{% for message in loop_messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|START_OF_TURN_TOKEN|><|USER_TOKEN|>' + content.strip() + '<|END_OF_TURN_TOKEN|>' }}{% elif message['role'] == 'assistant' %}{{ '<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>' + content.strip() + '<|END_OF_TURN_TOKEN|>' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>' }}{% endif %}"
308
+ },
309
+ {
310
+ "name": "tool_use",
311
+ "template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% set system_message = '## Task and Context\\nYou help people answer their questions and other requests interactively. You will be asked a very wide array of requests on all kinds of topics. You will be equipped with a wide range of search engines or similar tools to help you, which you use to research your answer. You should focus on serving the user\\'s needs as best you can, which will be wide-ranging.\\n\\n## Style Guide\\nUnless the user asks for a different style of answer, you should answer in full sentences, using proper grammar and spelling.' %}{% endif %}{{ '<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>' }}{{ '# Safety Preamble' }}{{ '\nThe instructions in this section override those in the task description and style guide sections. Don\\'t answer questions that are harmful or immoral.' }}{{ '\n\n# System Preamble' }}{{ '\n## Basic Rules' }}{{ '\nYou are a powerful conversational AI trained by Cohere to help people. You are augmented by a number of tools, and your job is to use and consume the output of these tools to best help the user. You will see a conversation history between yourself and a user, ending with an utterance from the user. You will then see a specific instruction instructing you what kind of response to generate. When you answer the user\\'s requests, you cite your sources in your answers, according to those instructions.' }}{{ '\n\n# User Preamble' }}{{ '\n' + system_message }}{{'\n\n## Available Tools\nHere is a list of tools that you have available to you:\n\n'}}{% for tool in tools %}{% if loop.index0 != 0 %}{{ '\n\n'}}{% endif %}{{'```python\ndef ' + tool.name + '('}}{% for param_name, param_fields in tool.parameter_definitions.items() %}{% if loop.index0 != 0 %}{{ ', '}}{% endif %}{{param_name}}: {% if not param_fields.required %}{{'Optional[' + param_fields.type + '] = None'}}{% else %}{{ param_fields.type }}{% endif %}{% endfor %}{{ ') -> List[Dict]:\n \"\"\"'}}{{ tool.description }}{% if tool.parameter_definitions|length != 0 %}{{ '\n\n Args:\n '}}{% for param_name, param_fields in tool.parameter_definitions.items() %}{% if loop.index0 != 0 %}{{ '\n ' }}{% endif %}{{ param_name + ' ('}}{% if not param_fields.required %}{{'Optional[' + param_fields.type + ']'}}{% else %}{{ param_fields.type }}{% endif %}{{ '): ' + param_fields.description }}{% endfor %}{% endif %}{{ '\n \"\"\"\n pass\n```' }}{% endfor %}{{ '<|END_OF_TURN_TOKEN|>'}}{% for message in loop_messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|START_OF_TURN_TOKEN|><|USER_TOKEN|>' + content.strip() + '<|END_OF_TURN_TOKEN|>' }}{% elif message['role'] == 'system' %}{{ '<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>' + content.strip() + '<|END_OF_TURN_TOKEN|>' }}{% elif message['role'] == 'assistant' %}{{ '<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>' + content.strip() + '<|END_OF_TURN_TOKEN|>' }}{% endif %}{% endfor %}{{'<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>Write \\'Action:\\' followed by a json-formatted list of actions that you want to perform in order to produce a good response to the user\\'s last input. You can use any of the supplied tools any number of times, but you should aim to execute the minimum number of necessary actions for the input. You should use the `directly-answer` tool if calling the other tools is unnecessary. The list of actions you want to call should be formatted as a list of json objects, for example:\n```json\n[\n {\n \"tool_name\": title of the tool in the specification,\n \"parameters\": a dict of parameters to input into the tool as they are defined in the specs, or {} if it takes no parameters\n }\n]```<|END_OF_TURN_TOKEN|>'}}{% if add_generation_prompt %}{{ '<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>' }}{% endif %}"
312
+ },
313
+ {
314
+ "name": "rag",
315
+ "template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% set system_message = '## Task and Context\\nYou help people answer their questions and other requests interactively. You will be asked a very wide array of requests on all kinds of topics. You will be equipped with a wide range of search engines or similar tools to help you, which you use to research your answer. You should focus on serving the user\\'s needs as best you can, which will be wide-ranging.\\n\\n## Style Guide\\nUnless the user asks for a different style of answer, you should answer in full sentences, using proper grammar and spelling.' %}{% endif %}{{ '<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>' }}{{ '# Safety Preamble' }}{{ '\nThe instructions in this section override those in the task description and style guide sections. Don\\'t answer questions that are harmful or immoral.' }}{{ '\n\n# System Preamble' }}{{ '\n## Basic Rules' }}{{ '\nYou are a powerful conversational AI trained by Cohere to help people. You are augmented by a number of tools, and your job is to use and consume the output of these tools to best help the user. You will see a conversation history between yourself and a user, ending with an utterance from the user. You will then see a specific instruction instructing you what kind of response to generate. When you answer the user\\'s requests, you cite your sources in your answers, according to those instructions.' }}{{ '\n\n# User Preamble' }}{{ '\n' + system_message }}{{ '<|END_OF_TURN_TOKEN|>'}}{% for message in loop_messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|START_OF_TURN_TOKEN|><|USER_TOKEN|>' + content.strip() + '<|END_OF_TURN_TOKEN|>' }}{% elif message['role'] == 'system' %}{{ '<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>' + content.strip() + '<|END_OF_TURN_TOKEN|>' }}{% elif message['role'] == 'assistant' %}{{ '<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>' + content.strip() + '<|END_OF_TURN_TOKEN|>' }}{% endif %}{% endfor %}{{ '<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>'}}{{ '<results>' }}{% for document in documents %}{{ '\nDocument: ' }}{{ loop.index0 }}\n{% for key, value in document.items() %}{{ key }}: {{value}}\n{% endfor %}{% endfor %}{{ '</results>'}}{{ '<|END_OF_TURN_TOKEN|><|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>' }}{{ 'Carefully perform the following instructions, in order, starting each with a new line.\n' }}{{ 'Firstly, Decide which of the retrieved documents are relevant to the user\\'s last input by writing \\'Relevant Documents:\\' followed by comma-separated list of document numbers. If none are relevant, you should instead write \\'None\\'.\n' }}{{ 'Secondly, Decide which of the retrieved documents contain facts that should be cited in a good answer to the user\\'s last input by writing \\'Cited Documents:\\' followed a comma-separated list of document numbers. If you dont want to cite any of them, you should instead write \\'None\\'.\n' }}{% if citation_mode=='accurate' %}{{ 'Thirdly, Write \\'Answer:\\' followed by a response to the user\\'s last input in high quality natural english. Use the retrieved documents to help you. Do not insert any citations or grounding markup.\n' }}{% endif %}{{ 'Finally, Write \\'Grounded answer:\\' followed by a response to the user\\'s last input in high quality natural english. Use the symbols <co: doc> and </co: doc> to indicate when a fact comes from a document in the search result, e.g <co: 0>my fact</co: 0> for a fact from document 0.' }}{{ '<|END_OF_TURN_TOKEN|>' }}{% if add_generation_prompt %}{{ '<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>' }}{% endif %}"
316
+ }
317
+ ],
318
+ "clean_up_tokenization_spaces": false,
319
+ "eos_token": "<|END_OF_TURN_TOKEN|>",
320
+ "legacy": true,
321
+ "merges_file": null,
322
+ "model_max_length": 1000000000000000019884624838656,
323
+ "pad_token": "<PAD>",
324
+ "sp_model_kwargs": {},
325
+ "spaces_between_special_tokens": false,
326
+ "tokenizer_class": "CohereTokenizer",
327
+ "unk_token": null,
328
+ "use_default_system_prompt": false,
329
+ "vocab_file": null
330
+ }