prince-canuma commited on
Commit
05b2789
1 Parent(s): d1ce7e3

Upload folder using huggingface_hub (#1)

Browse files

- 74723a3f7c9cc32f51c45f7dc2573f4639b0dd18ea68f1a23926fc2f41c4621d (622545d007229d5b5ae78e284c26dcab3d9318ec)
- d4698ad7a28d2ff457ecdc954160240452e2edd5f1bacdd0c9efdde9f8758e8c (e783de31b631449a7835cb6884e0f79977c3de9f)
- 3fb8365fcd664551e0c1063cf25e1183324a57085d7b2fc7b5eca453c7e90004 (ce1e6e5f5b3c4a84e7ab0b6eccabc034a1f21469)
- 69f6f160cc4436ea619a0e2eb363f6685f50e1f07acee0753dd34addc2406c08 (cf335c30f095f4dac9e64145915adce5549b903f)
- 6c8a67ba8efb0a99fa8ade0a42004c2c42943026751b91d489259335fbd216f9 (55a52729374e9ef15c9f2ca8e66c0d4ad36a966d)
- 8fe8d74f27f123c328a7dfae80a13cc7ac3eff888da1f72bc2bf5425fc89de42 (65c3b83e3f8424e2b2f6b9da66f07ed112fed84a)
- 3b6c1fae329383afe32880e2cff3523e4ffdd70622bff8eaea66d4e43143e302 (747058da5aa21534f158b1f68ecea6dacd0992d4)
- 914a58d3f3e6238522237a0d966b1cf3cee2531e4ac7e4352d4dace2bda9fd98 (14e904417954e8723cfae688508f4c4c0c495325)
- 5389187ee0c952e6aaeb37c98e687e78a5daad0f6036cc2d1d35dabaa39dbe2d (4c0700cd11e1503321bb1f0977c4b78f03128352)
- 28b2d54fb27cff9fa701e2ce7a0f446fa070c15dca7810a980044e3ed97f55fb (a60f90b1b18e47f0fa7bee33d6eda81f29c7bf75)
- 5d8e12fe5a80a884f52c437a2a1afe3a19a322df5269ba211ac0d837604d7d1d (366dcbf74305f9d3632c6c651edad35bbfdf7421)
- ccc3475d4e6686fc64e9a9d28f0436077059b2ebe30bb6146b0ed147cd5d485f (4b7e86f23fc97cb8f9d92be23ee312552a68c8ad)
- 88d52c6e510de151f14d3e003a27fa34be2d350dbca96c9820545701d126dc38 (7cb24081d2a5208f01bd16d2e538a2acb8d4019f)

README.md ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: apache-2.0
5
+ tags:
6
+ - pretrained
7
+ - moe
8
+ - mlx
9
+ pipeline_tag: text-generation
10
+ ---
11
+
12
+ # mlx-community/Qwen2-57B-A14B-8bit
13
+
14
+ The Model [mlx-community/Qwen2-57B-A14B-8bit](https://huggingface.co/mlx-community/Qwen2-57B-A14B-8bit) was converted to MLX format from [Qwen/Qwen2-57B-A14B](https://huggingface.co/Qwen/Qwen2-57B-A14B) using mlx-lm version **0.14.2**.
15
+
16
+ ## Use with mlx
17
+
18
+ ```bash
19
+ pip install mlx-lm
20
+ ```
21
+
22
+ ```python
23
+ from mlx_lm import load, generate
24
+
25
+ model, tokenizer = load("mlx-community/Qwen2-57B-A14B-8bit")
26
+ response = generate(model, tokenizer, prompt="hello", verbose=True)
27
+ ```
added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2MoeForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "decoder_sparse_step": 1,
8
+ "eos_token_id": 151643,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 3584,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 18944,
13
+ "max_position_embeddings": 131072,
14
+ "max_window_layers": 28,
15
+ "model_type": "qwen2_moe",
16
+ "moe_intermediate_size": 2560,
17
+ "norm_topk_prob": false,
18
+ "num_attention_heads": 28,
19
+ "num_experts": 64,
20
+ "num_experts_per_tok": 8,
21
+ "num_hidden_layers": 28,
22
+ "num_key_value_heads": 4,
23
+ "output_router_logits": false,
24
+ "quantization": {
25
+ "group_size": 64,
26
+ "bits": 8
27
+ },
28
+ "rms_norm_eps": 1e-06,
29
+ "rope_theta": 1000000.0,
30
+ "router_aux_loss_coef": 0.001,
31
+ "shared_expert_intermediate_size": 20480,
32
+ "sliding_window": 131072,
33
+ "tie_word_embeddings": false,
34
+ "torch_dtype": "bfloat16",
35
+ "transformers_version": "4.40.1",
36
+ "use_cache": true,
37
+ "use_sliding_window": false,
38
+ "vocab_size": 151936
39
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model-00001-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78335e3e6a0456156bd5d4b874528186aac34a9e6d458372ed895a37f8fd600a
3
+ size 4884308032
model-00002-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8abafbe8b00fef865ed44fdb74b5cdf492ec49c3984f1d1ff24b4c92a96f3e4
3
+ size 4898188502
model-00003-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c1896c209372db6c3f0b1abc41887242e171c004c37bd6946c532ce535d3c8e
3
+ size 4898188506
model-00004-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3de881e68e17125d57677f13be0034ce6a2a25a995be236e9859a6c492edb565
3
+ size 5163621926
model-00005-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c52b971f3a5e4b3a285a7c896923bb4d4c54090dc559c1a84f5723ce432e61e
3
+ size 4898188608
model-00006-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e29e44f8c0e11af04e30e4dc09d5da433698d02006bdc28661cc58630a70d8f7
3
+ size 4898188619
model-00007-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efa6a0a7ee8cd7c3a8ce0a10f484c3c7bd10498adfc479533aadca3e2793d392
3
+ size 5163622045
model-00008-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8727cbb04d8572cd52b7aa36defeaa4090d3746c9b961d26c6498e835de7ae1
3
+ size 4898188575
model-00009-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:586357ff7782f8a559c77ebb7028ac9c7b3d46fae317e57f4e2c390dd88ea50a
3
+ size 4898188571
model-00010-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44f0362b7ed4c3f8e9e5888d4ba4114d31408c00ed3a27c66fc83886789d31f1
3
+ size 5163622027
model-00011-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c711b2a2efac621ec01d64b015dc97d1672934bc65e9f5b18236268bf081effe
3
+ size 4898188591
model-00012-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e39057a1a5806f1bdbc5dfd4a985e77703f243b3b5c80455cbcbe593b4b6e43
3
+ size 4898188561
model-00013-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55367daee105fa4bc70b7d3f277f9a4ea5e3839882cebd9ee82bb0747054bcce
3
+ size 1436466320
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|endoftext|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "pad_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ }
20
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
+ "clean_up_tokenization_spaces": false,
36
+ "eos_token": "<|endoftext|>",
37
+ "errors": "replace",
38
+ "model_max_length": 131072,
39
+ "pad_token": "<|endoftext|>",
40
+ "split_special_tokens": false,
41
+ "tokenizer_class": "Qwen2Tokenizer",
42
+ "unk_token": null
43
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff