prince-canuma commited on
Commit
9b12f3e
1 Parent(s): cfefbd4

Upload folder using huggingface_hub (#1)

Browse files

- b49b19830c3e6eabe73c9d1ad24bcfb7c42605fb187e5f1c68cd1db12ec37def (52c257b00e629c2cc1563c8d51d931bf570b50b3)
- 872feaf4cc5438fde5e967a7e7181b82bd80609531c0d7d93ba064f7186e5b40 (fc8903c9eeabb199e2806da40e5bb5145c4d4307)
- c3eebcfb6154da588b09006d871db513741f65bd58802458d37e0086c8c51e45 (578dcb7b8f8e0524b490dc162087e875c63f288c)
- add0382ee8ac59fcd38e16d2834c7bb5e0638ca5e258702836be80de2036bb20 (8162d040e71a6711be529d1f4cb77e5182027c6f)
- c815939ea5d5bf475838a60dab19a7750c6203978d41928217eb20e8ed42abe6 (26bf9f50867029540fb4203a91a688ae1d9c36f6)
- bae2499279ff3b5c0f4245da05592fa5dc3ad07e057d3fd7714087bf4f1f2ec0 (908f7f7f0ae2bc8e65f02436329e07f993e0d16f)
- 7e4fcafecd52691a71c67b198742586ed7e2f0f5ec839118c4c6ce91987fa89d (d353f90508eee4f2bd0fa39ec1255f328c952943)
- 1d15e352d7633a70b8c19d1a7924422c09df6b84d93efbe4abd6b6d80c01e4fe (4a42b06142b41988b8557def17e25ba89edd1a92)
- 6b39028b15d1b65e8d86074ae567c1f0f1b4f83545b559d9365583554c1f12b2 (7552902e8aef3627d1fa927c5db1622ac196d30d)
- 3d44b3fa3d910b07e810f497e573e1e26e5b773e8a8148a66ccbaed7721435bf (4ae5def012a70640f39ff00e005df9756e4733ca)
- 64017e9c65220213803a331244a9a06f9548913618c144218dcafe7aafb41276 (bf15603cc188e1f8228bb8344f433e91d49f5229)
- 181ed9751a72824d77f5dce31db78fe51b7edf6c47303b3731e582780bc2f302 (789b74f33fa7ab0b396264ec3d3fdb50396de32f)
- 53b9435a66f135e8e1c282be019c17a987dfe38c4fe417de2faa516163eb6c26 (e6ea3e545ce2df3c4b933f53e6f71419f016d8ba)

README.md ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: apache-2.0
5
+ tags:
6
+ - chat
7
+ - mlx
8
+ pipeline_tag: text-generation
9
+ ---
10
+
11
+ # mlx-community/Qwen2-57B-A14B-Instruct-8bit
12
+
13
+ The Model [mlx-community/Qwen2-57B-A14B-Instruct-8bit](https://huggingface.co/mlx-community/Qwen2-57B-A14B-Instruct-8bit) was converted to MLX format from [Qwen/Qwen2-57B-A14B-Instruct](https://huggingface.co/Qwen/Qwen2-57B-A14B-Instruct) using mlx-lm version **0.14.2**.
14
+
15
+ ## Use with mlx
16
+
17
+ ```bash
18
+ pip install mlx-lm
19
+ ```
20
+
21
+ ```python
22
+ from mlx_lm import load, generate
23
+
24
+ model, tokenizer = load("mlx-community/Qwen2-57B-A14B-Instruct-8bit")
25
+ response = generate(model, tokenizer, prompt="hello", verbose=True)
26
+ ```
added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2MoeForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "decoder_sparse_step": 1,
8
+ "eos_token_id": 151643,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 3584,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 18944,
13
+ "max_position_embeddings": 32768,
14
+ "max_window_layers": 28,
15
+ "model_type": "qwen2_moe",
16
+ "moe_intermediate_size": 2560,
17
+ "norm_topk_prob": false,
18
+ "num_attention_heads": 28,
19
+ "num_experts": 64,
20
+ "num_experts_per_tok": 8,
21
+ "num_hidden_layers": 28,
22
+ "num_key_value_heads": 4,
23
+ "output_router_logits": false,
24
+ "quantization": {
25
+ "group_size": 64,
26
+ "bits": 8
27
+ },
28
+ "rms_norm_eps": 1e-06,
29
+ "rope_theta": 1000000.0,
30
+ "router_aux_loss_coef": 0.001,
31
+ "shared_expert_intermediate_size": 20480,
32
+ "sliding_window": 65536,
33
+ "tie_word_embeddings": false,
34
+ "torch_dtype": "bfloat16",
35
+ "transformers_version": "4.40.1",
36
+ "use_cache": true,
37
+ "use_sliding_window": false,
38
+ "vocab_size": 151936
39
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model-00001-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a207246dafd8948d7e8b1ae6a373a5f8b3be851808d2f568145d572d84919b7
3
+ size 4884308032
model-00002-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aacd596b61729d92c9b4b86e0e3fc586b82a32e31b6b3afcd71c90b5c0946926
3
+ size 4898188502
model-00003-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a79d8842981d6c679666f02a978ad4bafee13fd4e8a773dfc0c506e0d550e7d0
3
+ size 4898188506
model-00004-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24a8ad4d3bd0faf530baa32352abd5386870044fdc12223b67eb7e29612e7a42
3
+ size 5163621926
model-00005-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37ff376d2cd635fbad5c17b3ee766422b207ec9fe9dd6905d016238f7e4a4cb4
3
+ size 4898188608
model-00006-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a94a584c9167ceb454f2dcc102c1feaff491b4b3afa07c94f21284f03865a464
3
+ size 4898188619
model-00007-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfd754910c6fa4aee4e8fc493b279c28eca6278818a48acfe71c09c8e9ff74d2
3
+ size 5163622045
model-00008-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3efde7bc39e10cbe3bb7c41be63db20f9e4ef0d60f155109ce78d0bf2bacdd5e
3
+ size 4898188575
model-00009-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7533b1d1ee3b0a403b741fb1756d6d34c5d1a66e50d86b7942d5c25fcd62eb6
3
+ size 4898188571
model-00010-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc1c5301aa81d77254300dcab80af42c8cb167666ca10a81540930938e4f9105
3
+ size 5163622027
model-00011-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78fca7206cf59d24a878e2e743ccbaa5e0f7eab48d68a00a9c6ed5c0b7eebefc
3
+ size 4898188591
model-00012-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5e0293cb98b76be88a571eb7430f1a69bb24a673a6d8f2d2c7a63966232389f
3
+ size 4898188561
model-00013-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:244607f0a80608a9c375266a15af8fdf20495fb388396fb4ad9077764072d5d7
3
+ size 1436466320
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|im_end|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "pad_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ }
20
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
+ "clean_up_tokenization_spaces": false,
36
+ "eos_token": "<|im_end|>",
37
+ "errors": "replace",
38
+ "model_max_length": 65536,
39
+ "pad_token": "<|endoftext|>",
40
+ "split_special_tokens": false,
41
+ "tokenizer_class": "Qwen2Tokenizer",
42
+ "unk_token": null
43
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff