prince-canuma commited on
Commit
ee989d7
1 Parent(s): e3e909c

Upload folder using huggingface_hub (#1)

Browse files

- 3535051602e7215515920a34b2b88a9899a6623b6f3cb7633e832e9808c1cb4a (e0911cbc6448c12e9104daf6748cd51ab586c19d)
- a0664c76bf69138620f63c2a791325d197bc50540130434b2966ab839b62435d (34e1b89902b509f14d6a3d4684f4939fbd57280f)
- 5b3a493d6b92ce37d303fa5e6c53af2f6df8c158c0c7209f139d7516f3c33760 (6851e3b4adbabdde84d8125c3547572e239df061)
- 10b4ae26f45d6454857fae31622c95bd911e3932b12f9259991a5fdd8e7eccaf (1a84e16d37de5fbb8e9dd299c851173035862c77)
- 8a2482a26a99174f25a3a4c54dfd1d2cf0274869ea103d85c6ee6dd16cadde31 (089cf98e7f52979dc503b5764690ffe720351ba8)
- 537d288b469f898f3d939f809eaaed5c180478feb639eec06210f902a531749a (441ba9eb5d6812b7bbc216e40d927c98e1a78d0a)
- 32d8202e033f201f4f78b6024452b2ed325ec64127566b324e3b9ae3172c6cf9 (44c44b7f5e30bc49fd0550b889d2c4790b6d0390)
- cfbb1783ee56d636113817e8b4a3459dbda11b161bd204c81c85576b363be15f (2e5f9930503d1df15d02f0867a82c1cdc4250fcb)

README.md ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - mlx
5
+ ---
6
+
7
+ # mlx-community/Yi-1.5-34B-8bit
8
+
9
+ The Model [mlx-community/Yi-1.5-34B-8bit](https://huggingface.co/mlx-community/Yi-1.5-34B-8bit) was converted to MLX format from [01-ai/Yi-1.5-34B](https://huggingface.co/01-ai/Yi-1.5-34B) using mlx-lm version **0.13.0**.
10
+
11
+ ## Use with mlx
12
+
13
+ ```bash
14
+ pip install mlx-lm
15
+ ```
16
+
17
+ ```python
18
+ from mlx_lm import load, generate
19
+
20
+ model, tokenizer = load("mlx-community/Yi-1.5-34B-8bit")
21
+ response = generate(model, tokenizer, prompt="hello", verbose=True)
22
+ ```
config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 7168,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 20480,
13
+ "max_position_embeddings": 4096,
14
+ "model_type": "llama",
15
+ "num_attention_heads": 56,
16
+ "num_hidden_layers": 60,
17
+ "num_key_value_heads": 8,
18
+ "pad_token_id": 0,
19
+ "pretraining_tp": 1,
20
+ "quantization": {
21
+ "group_size": 64,
22
+ "bits": 8
23
+ },
24
+ "rms_norm_eps": 1e-06,
25
+ "rope_scaling": null,
26
+ "rope_theta": 5000000.0,
27
+ "tie_word_embeddings": false,
28
+ "torch_dtype": "bfloat16",
29
+ "transformers_version": "4.37.2",
30
+ "use_cache": true,
31
+ "vocab_size": 64000
32
+ }
model-00001-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49460c134e9b3c52086c43df4ac5a271b1ed41dd8d05b4e2b491f0fbe51630e3
3
+ size 5354117230
model-00002-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1713aca52d0de5a0b28a944c13110c9e5745b03d732f05ce2ce3184ea8e7a48a
3
+ size 5334650055
model-00003-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6aeb5c2dcb14b53ac3ce649709e0cd5985eb5f415d21f5c55379583896758e76
3
+ size 5334650069
model-00004-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:178292b652d9931b45f4308ac1822caac63ef4fb833bbda27c58ccf6ec46e3fc
3
+ size 5334650063
model-00005-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:673549c660f283b271604c23dfd4e34e9b1fabe604507c6e9ac861d87686fc46
3
+ size 5334650069
model-00006-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f3405f77543c732bd1bd6ab406557b3991ad9151f076b38361ddace13ec8577
3
+ size 5334650073
model-00007-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0ccb2c351644ceb2b96e53b4e311e6b51f739b7c7eb59262552276d6987f020
3
+ size 4511829028
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:386c49cf943d71aa110361135338c50e38beeff0a66593480421f37b319e1a39
3
+ size 1033105
tokenizer_config.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<|startoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "<|endoftext|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ }
29
+ },
30
+ "bos_token": "<|startoftext|>",
31
+ "clean_up_tokenization_spaces": false,
32
+ "eos_token": "<|endoftext|>",
33
+ "legacy": true,
34
+ "model_max_length": 4096,
35
+ "pad_token": "<unk>",
36
+ "sp_model_kwargs": {},
37
+ "tokenizer_class": "LlamaTokenizer",
38
+ "unk_token": "<unk>",
39
+ "use_default_system_prompt": false
40
+ }